Skip to main content

Showing 1–50 of 181 results for author: Zhong, M

Searching in archive cs. Search in all archives.
.
  1. arXiv:2604.00835  [pdf, ps, other

    cs.CL

    Agentic Tool Use in Large Language Models

    Authors: Jinchao Hu, Meizhi Zhong, Kehai Chen, Xuefeng Bai, Min Zhang

    Abstract: Large language models are increasingly being deployed as autonomous agents yet their real world effectiveness depends on reliable tools for information retrieval, computation and external action. Existing studies remain fragmented across tasks, tool types, and training settings, lacking a unified view of how tool-use methods differ and evolve. This paper organizes the literature into three paradig… ▽ More

    Submitted 1 April, 2026; originally announced April 2026.

  2. arXiv:2603.11076  [pdf, ps, other

    cs.AI cs.SE

    DIVE: Scaling Diversity in Agentic Task Synthesis for Generalizable Tool Use

    Authors: Aili Chen, Chi Zhang, Junteng Liu, Jiangjie Chen, Chengyu Du, Yunji Li, Ming Zhong, Qin Wang, Zhengmao Zhu, Jiayuan Song, Ke Ji, Junxian He, Pengyu Zhao, Yanghua Xiao

    Abstract: Recent work synthesizes agentic tasks for post-training tool-using LLMs, yet robust generalization under shifts in tasks and toolsets remains an open challenge. We trace this brittleness to insufficient diversity in synthesized tasks. Scaling diversity is difficult because training requires tasks to remain executable and verifiable, while generalization demands coverage of diverse tool types, tool… ▽ More

    Submitted 10 March, 2026; originally announced March 2026.

  3. arXiv:2603.09865  [pdf, ps, other

    cs.LG

    GAST: Gradient-aligned Sparse Tuning of Large Language Models with Data-layer Selection

    Authors: Kai Yao, Zhenghan Song, Kaixin Wu, Mingjie Zhong, Danzhao Cheng, Zhaorui Tan, Yixin Ji, Penglei Gao

    Abstract: Parameter-Efficient Fine-Tuning (PEFT) has become a key strategy for adapting large language models, with recent advances in sparse tuning reducing overhead by selectively updating key parameters or subsets of data. Existing approaches generally focus on two distinct paradigms: layer-selective methods aiming to fine-tune critical layers to minimize computational load, and data-selective methods ai… ▽ More

    Submitted 10 March, 2026; originally announced March 2026.

  4. arXiv:2602.22808  [pdf, ps, other

    cs.AI

    MiroFlow: Towards High-Performance and Robust Open-Source Agent Framework for General Deep Research Tasks

    Authors: Shiqian Su, Sen Xing, Xuan Dong, Muyan Zhong, Bin Wang, Xizhou Zhu, Yuntao Chen, Wenhai Wang, Yue Deng, Pengxiang Zhu, Ziyuan Liu, Tiantong Li, Jiaheng Yu, Zhe Chen, Lidong Bing, Jifeng Dai

    Abstract: Despite the remarkable progress of large language models (LLMs), the capabilities of standalone LLMs have begun to plateau when tackling real-world, complex tasks that require interaction with external tools and dynamic environments. Although recent agent frameworks aim to enhance model autonomy through tool integration and external interaction, they still suffer from naive workflows, unstable per… ▽ More

    Submitted 26 February, 2026; originally announced February 2026.

  5. arXiv:2602.18511  [pdf, ps, other

    cs.PL cs.AI

    Beyond Pass-by-Pass Optimization: Intent-Driven IR Optimization with Large Language Models

    Authors: Lei Qiu, Zi Yang, Fang Lyu, Ming Zhong, Huimin Cui, Xiaobing Feng

    Abstract: Modern compilers optimize programs through a sequence of modular passes over intermediate representations (IR). While this pass-by-pass paradigm offers engineering benefits, it suffers from a pass coordination problem: locally beneficial transformations may block more profitable optimizations in later stages. This limitation stems from the lack of an explicit notion of optimization intent, defined… ▽ More

    Submitted 19 February, 2026; originally announced February 2026.

  6. arXiv:2602.14043  [pdf, ps, other

    cs.SI cs.AI

    Beyond Static Snapshots: Dynamic Modeling and Forecasting of Group-Level Value Evolution with Large Language Models

    Authors: Qiankun Pi, Guixin Su, Jinliang Li, Mayi Xu, Xin Miao, Jiawei Jiang, Ming Zhong, Tieyun Qian

    Abstract: Social simulation is critical for mining complex social dynamics and supporting data-driven decision making. LLM-based methods have emerged as powerful tools for this task by leveraging human-like social questionnaire responses to model group behaviors. Existing LLM-based approaches predominantly focus on group-level values at discrete time points, treating them as static snapshots rather than dyn… ▽ More

    Submitted 15 February, 2026; originally announced February 2026.

  7. arXiv:2602.07276  [pdf, ps, other

    cs.AI cs.CL cs.LG

    Steer2Adapt: Dynamically Composing Steering Vectors Elicits Efficient Adaptation of LLMs

    Authors: Pengrui Han, Xueqiang Xu, Keyang Xuan, Peiyang Song, Siru Ouyang, Runchu Tian, Yuqing Jiang, Cheng Qian, Pengcheng Jiang, Jiashuo Sun, Junxia Cui, Ming Zhong, Ge Liu, Jiawei Han, Jiaxuan You

    Abstract: Activation steering has emerged as a promising approach for efficiently adapting large language models (LLMs) to downstream behaviors. However, most existing steering methods rely on a single static direction per task or concept, making them inflexible under task variation and inadequate for complex tasks that require multiple coordinated capabilities. To address this limitation, we propose STEER2… ▽ More

    Submitted 6 February, 2026; originally announced February 2026.

  8. arXiv:2602.03689  [pdf, ps, other

    cs.CL cs.AI

    Rethinking the Reranker: Boundary-Aware Evidence Selection for Robust Retrieval-Augmented Generation

    Authors: Jiashuo Sun, Pengcheng Jiang, Saizhuo Wang, Jiajun Fan, Heng Wang, Siru Ouyang, Ming Zhong, Yizhu Jiao, Chengsong Huang, Xueqiang Xu, Pengrui Han, Peiran Li, Jiaxin Huang, Ge Liu, Heng Ji, Jiawei Han

    Abstract: Retrieval-Augmented Generation (RAG) systems remain brittle under realistic retrieval noise, even when the required evidence appears in the top-K results. A key reason is that retrievers and rerankers optimize solely for relevance, often selecting either trivial, answer-revealing passages or evidence that lacks the critical information required to answer the question, without considering whether t… ▽ More

    Submitted 3 February, 2026; originally announced February 2026.

    Comments: 19 pages, 8 tables, 5 figures

  9. arXiv:2602.00966  [pdf, ps, other

    cs.MA

    Symphony-Coord: Emergent Coordination in Decentralized Agent Systems

    Authors: Zhaoyang Guan, Huixi Cao, Ming Zhong, Eric Yang, Lynn Ai, Yongxin Ni, Bill Shi

    Abstract: Multi-agent large language model systems can tackle complex multi-step tasks by decomposing work and coordinating specialized behaviors. However, current coordination mechanisms typically rely on statically assigned roles and centralized controllers. As agent pools and task distributions evolve, these design choices lead to inefficient routing, poor adaptability, and fragile fault recovery capabil… ▽ More

    Submitted 31 January, 2026; originally announced February 2026.

    Comments: 41 pages,15 figures

  10. arXiv:2601.21919  [pdf, ps, other

    cs.AI cs.CL

    Self-Compression of Chain-of-Thought via Multi-Agent Reinforcement Learning

    Authors: Yiqun Chen, Jinyuan Feng, Wei Yang, Meizhi Zhong, Zhengliang Shi, Rui Li, Xiaochi Wei, Yan Gao, Yi Wu, Yao Hu, Zhiqiang Pu, Jiaxin Mao

    Abstract: The inference overhead induced by redundant reasoning undermines the interactive experience and severely bottlenecks the deployment of Large Reasoning Models. Existing reinforcement learning (RL)-based solutions tackle this problem by coupling a length penalty with outcome-based rewards. This simplistic reward weighting struggles to reconcile brevity with accuracy, as enforcing brevity may comprom… ▽ More

    Submitted 29 January, 2026; originally announced January 2026.

  11. arXiv:2601.21916  [pdf, ps, other

    cs.AI cs.CL cs.IR

    JADE: Bridging the Strategic-Operational Gap in Dynamic Agentic RAG

    Authors: Yiqun Chen, Erhan Zhang, Tianyi Hu, Shijie Wang, Zixuan Yang, Meizhi Zhong, Xiaochi Wei, Yan Gao, Yi Wu, Yao Hu, Jiaxin Mao

    Abstract: The evolution of Retrieval-Augmented Generation (RAG) has shifted from static retrieval pipelines to dynamic, agentic workflows where a central planner orchestrates multi-turn reasoning. However, existing paradigms face a critical dichotomy: they either optimize modules jointly within rigid, fixed-graph architectures, or empower dynamic planning while treating executors as frozen, black-box tools.… ▽ More

    Submitted 29 January, 2026; originally announced January 2026.

  12. arXiv:2601.21239  [pdf, ps, other

    cs.AI

    TIDE: Tuning-Integrated Dynamic Evolution for LLM-Based Automated Heuristic Design

    Authors: Chentong Chen, Mengyuan Zhong, Ye Fan, Jialong Shi, Jianyong Sun

    Abstract: Although Large Language Models have advanced Automated Heuristic Design, treating algorithm evolution as a monolithic text generation task overlooks the coupling between discrete algorithmic structures and continuous numerical parameters. Consequently, existing methods often discard promising algorithms due to uncalibrated constants and suffer from premature convergence resulting from simple simil… ▽ More

    Submitted 7 February, 2026; v1 submitted 28 January, 2026; originally announced January 2026.

  13. arXiv:2601.08223  [pdf, ps, other

    cs.CR cs.AI

    DNF: Dual-Layer Nested Fingerprinting for Large Language Model Intellectual Property Protection

    Authors: Zhenhua Xu, Yiran Zhao, Mengting Zhong, Dezhang Kong, Changting Lin, Tong Qiao, Meng Han

    Abstract: The rapid growth of large language models raises pressing concerns about intellectual property protection under black-box deployment. Existing backdoor-based fingerprints either rely on rare tokens -- leading to high-perplexity inputs susceptible to filtering -- or use fixed trigger-response mappings that are brittle to leakage and post-hoc adaptation. We propose \textsc{Dual-Layer Nested Fingerpr… ▽ More

    Submitted 21 January, 2026; v1 submitted 13 January, 2026; originally announced January 2026.

    Comments: Accepted by ICASSP2026

  14. arXiv:2601.06920  [pdf, ps, other

    cs.NE cs.MA

    Calibrating Agent-Based Financial Markets Simulators with Pretrainable Automatic Posterior Transformation-Based Surrogates

    Authors: Boquan Jiang, Zhenhua Yang, Chenkai Wang, Muyao Zhong, Heping Fang, Peng Yang

    Abstract: Calibrating Agent-Based Models (ABMs) is an important optimization problem for simulating the complex social systems, where the goal is to identify the optimal parameter of a given ABM by minimizing the discrepancy between the simulated data and the real-world observations. Unfortunately, it suffers from the extensive computational costs of iterative evaluations, which involves the expensive simul… ▽ More

    Submitted 11 January, 2026; originally announced January 2026.

    Comments: 32 pages, 4 figures

  15. arXiv:2512.16301  [pdf, ps, other

    cs.AI cs.CL

    Adaptation of Agentic AI: A Survey of Post-Training, Memory, and Skills

    Authors: Pengcheng Jiang, Jiacheng Lin, Zhiyi Shi, Zifeng Wang, Luxi He, Yichen Wu, Ming Zhong, Peiyang Song, Qizheng Zhang, Heng Wang, Xueqiang Xu, Hanwen Xu, Pengrui Han, Dylan Zhang, Jiashuo Sun, Chaoqi Yang, Kun Qian, Tian Wang, Changran Hu, Manling Li, Quanzheng Li, Hao Peng, Sheng Wang, Jingbo Shang, Chao Zhang , et al. (9 additional authors not shown)

    Abstract: Large language model (LLM) agents are moving beyond prompting alone. ChatGPT marked the rise of general-purpose LLM assistants, DeepSeek showed that on-policy reinforcement learning with verifiable rewards can improve reasoning and tool use, and OpenClaw highlights a newer direction in which agents accumulate persistent memory and reusable skills. Yet the research landscape remains fragmented acro… ▽ More

    Submitted 9 March, 2026; v1 submitted 18 December, 2025; originally announced December 2025.

  16. arXiv:2512.04515  [pdf, ps, other

    cs.CV

    EgoLCD: Egocentric Video Generation with Long Context Diffusion

    Authors: Liuzhou Zhang, Jiarui Ye, Yuanlei Wang, Ming Zhong, Mingju Cao, Wanke Xia, Bowen Zeng, Zeyu Zhang, Hao Tang

    Abstract: Generating long, coherent egocentric videos is difficult, as hand-object interactions and procedural tasks require reliable long-term memory. Existing autoregressive models suffer from content drift, where object identity and scene semantics degrade over time. To address this challenge, we introduce EgoLCD, an end-to-end framework for egocentric long-context video generation that treats long video… ▽ More

    Submitted 4 December, 2025; originally announced December 2025.

  17. arXiv:2511.19956  [pdf, ps, other

    cs.LG cs.IT

    Prompt Fairness: Sub-group Disparities in LLMs

    Authors: Meiyu Zhong, Noel Teku, Ravi Tandon

    Abstract: Large Language Models (LLMs), though shown to be effective in many applications, can vary significantly in their response quality. In this paper, we investigate this problem of prompt fairness: specifically, the phrasing of a prompt by different users/styles, despite the same question being asked in principle, may elicit different responses from an LLM. To quantify this disparity, we propose to us… ▽ More

    Submitted 25 November, 2025; originally announced November 2025.

  18. arXiv:2511.18121  [pdf, ps, other

    cs.CV cs.AI

    VCU-Bridge: Hierarchical Visual Connotation Understanding via Semantic Bridging

    Authors: Ming Zhong, Yuanlei Wang, Liuzhou Zhang, Arctanx An, Renrui Zhang, Hao Liang, Ming Lu, Ying Shen, Wentao Zhang

    Abstract: While Multimodal Large Language Models (MLLMs) excel on benchmarks, their processing paradigm differs from the human ability to integrate visual information. Unlike humans who naturally bridge details and high-level concepts, models tend to treat these elements in isolation. Prevailing evaluation protocols often decouple low-level perception from high-level reasoning, overlooking their semantic an… ▽ More

    Submitted 22 November, 2025; originally announced November 2025.

  19. arXiv:2511.14469  [pdf, ps, other

    cs.CV

    CompEvent: Complex-valued Event-RGB Fusion for Low-light Video Enhancement and Deblurring

    Authors: Mingchen Zhong, Xin Lu, Dong Li, Senyan Xu, Ruixuan Jiang, Xueyang Fu, Baocai Yin

    Abstract: Low-light video deblurring poses significant challenges in applications like nighttime surveillance and autonomous driving due to dim lighting and long exposures. While event cameras offer potential solutions with superior low-light sensitivity and high temporal resolution, existing fusion methods typically employ staged strategies, limiting their effectiveness against combined low-light and motio… ▽ More

    Submitted 6 February, 2026; v1 submitted 18 November, 2025; originally announced November 2025.

  20. arXiv:2511.11793  [pdf, ps, other

    cs.CL

    MiroThinker: Pushing the Performance Boundaries of Open-Source Research Agents via Model, Context, and Interactive Scaling

    Authors: MiroMind Team, Song Bai, Lidong Bing, Carson Chen, Guanzheng Chen, Yuntao Chen, Zhe Chen, Ziyi Chen, Jifeng Dai, Xuan Dong, Wenhan Dou, Yue Deng, Yunjie Fu, Junqi Ge, Chenxia Han, Tammy Huang, Zhenhang Huang, Jerry Jiao, Shilei Jiang, Tianyu Jiao, Xiaoqi Jian, Lei Lei, Ruilin Li, Ryan Luo, Tiantong Li , et al. (30 additional authors not shown)

    Abstract: We present MiroThinker v1.0, an open-source research agent designed to advance tool-augmented reasoning and information-seeking capabilities. Unlike previous agents that only scale up model size or context length, MiroThinker explores interaction scaling at the model level, systematically training the model to handle deeper and more frequent agent-environment interactions as a third dimension of p… ▽ More

    Submitted 18 November, 2025; v1 submitted 14 November, 2025; originally announced November 2025.

    Comments: Technical Report

  21. arXiv:2511.10921  [pdf, ps, other

    quant-ph cs.AR

    A Compilation Framework for Quantum Circuits with Mid-Circuit Measurement Error Awareness

    Authors: Ming Zhong, Zhemin Zhang, Xiangyu Ren, Chenghong Zhu, Siyuan Niu, Zhiding Liang

    Abstract: Mid-circuit measurement (MCM) provides the capability for qubit reuse and dynamic control in quantum processors, enabling more resource-efficient algorithms and supporting error-correction procedures. However, MCM introduces several sources of error, including measurement-induced crosstalk, idling-qubit decoherence, and reset infidelity, and these errors exhibit pronounced qubit-dependent variabil… ▽ More

    Submitted 13 November, 2025; originally announced November 2025.

    Comments: 8 pages, 7 figures

    ACM Class: C.1.3; D.3.4

  22. arXiv:2511.05854  [pdf, ps, other

    cs.AI

    Can a Small Model Learn to Look Before It Leaps? Dynamic Learning and Proactive Correction for Hallucination Detection

    Authors: Zepeng Bao, Shen Zhou, Qiankun Pi, Jianhao Chen, Mayi Xu, Ming Zhong, Yuanyuan Zhu, Tieyun Qian

    Abstract: Hallucination in large language models (LLMs) remains a critical barrier to their safe deployment. For hallucination detection to be practical in real-world scenarios, the use of efficient small models is essential to ensure low latency and minimal resource consumption. However, existing methods rely on fixed verification strategies, where simply tuning small models to mimic fixed verification tra… ▽ More

    Submitted 4 March, 2026; v1 submitted 8 November, 2025; originally announced November 2025.

  23. arXiv:2510.17847  [pdf, ps, other

    cs.CV

    CoIDO: Efficient Data Selection for Visual Instruction Tuning via Coupled Importance-Diversity Optimization

    Authors: Yichen Yan, Ming Zhong, Qi Zhu, Xiaoling Gu, Jinpeng Chen, Huan Li

    Abstract: Multimodal large language models (MLLMs) rely heavily on instruction tuning to align vision and language capabilities, yet the computational cost of training on large-scale datasets remains a major bottleneck. Existing data selection methods aim to mitigate this by selecting important and diverse subsets, but they often suffer from two critical drawbacks: high computational overhead from processin… ▽ More

    Submitted 11 October, 2025; originally announced October 2025.

    Comments: 22 pages, 8 figures, 39th Conference on Neural Information Processing Systems (NeurIPS 2025)

  24. arXiv:2510.16816  [pdf, ps, other

    cs.LG cs.AI math-ph physics.comp-ph

    Efficient High-Accuracy PDEs Solver with the Linear Attention Neural Operator

    Authors: Ming Zhong, Zhenya Yan

    Abstract: Neural operators offer a powerful data-driven framework for learning mappings between function spaces, in which the transformer-based neural operator architecture faces a fundamental scalability-accuracy trade-off: softmax attention provides excellent fidelity but incurs quadratic complexity $\mathcal{O}(N^2 d)$ in the number of mesh points $N$ and hidden dimension $d$, while linear attention vari… ▽ More

    Submitted 19 October, 2025; originally announced October 2025.

    Comments: 31 pages, 8 figures

  25. TaskAudit: Detecting Functiona11ity Errors in Mobile Apps via Agentic Task Execution

    Authors: Mingyuan Zhong, Xia Chen, Davin Win Kyi, Chen Li, James Fogarty, Jacob O. Wobbrock

    Abstract: Accessibility checkers are tools in support of accessible app development, and their use is encouraged by accessibility best practices. However, most current checkers evaluate static or mechanically-generated contexts, failing to capture common accessibility errors impacting mobile app functionality. In this work, we define functiona11ity errors as accessibility barriers that only manifest through… ▽ More

    Submitted 4 February, 2026; v1 submitted 14 October, 2025; originally announced October 2025.

    Comments: CHI 2026

    ACM Class: H.5.2

  26. SusBench: An Online Benchmark for Evaluating Dark Pattern Susceptibility of Computer-Use Agents

    Authors: Longjie Guo, Chenjie Yuan, Mingyuan Zhong, Robert Wolfe, Ruican Zhong, Yue Xu, Bingbing Wen, Hua Shen, Lucy Lu Wang, Alexis Hiniker

    Abstract: As LLM-based computer-use agents (CUAs) begin to autonomously interact with real-world interfaces, understanding their vulnerability to manipulative interface designs becomes increasingly critical. We introduce SusBench, an online benchmark for evaluating the susceptibility of CUAs to UI dark patterns, designs that aim to manipulate or deceive users into taking unintentional actions. Drawing nine… ▽ More

    Submitted 23 February, 2026; v1 submitted 13 October, 2025; originally announced October 2025.

    Comments: Accepted as a full paper to IUI 2026

  27. arXiv:2510.09942  [pdf, ps, other

    cs.LG cs.AI cs.IT

    Conformal Sparsification for Bandwidth-Efficient Edge-Cloud Speculative Decoding

    Authors: Payel Bhattacharjee, Fengwei Tian, Meiyu Zhong, Guangyi Zhang, Osvaldo Simeone, Ravi Tandon

    Abstract: Edge-cloud speculative decoding (SD) accelerates inference by having a cloud-based large language model (LLM) that verifies draft tokens generated by a resource-constrained small language model (SLM) at the edge. A central bottleneck is the limited bandwidth of the edge-cloud link, which necessitates efficient compression of draft token distributions. We first derive an information-theoretic bound… ▽ More

    Submitted 10 October, 2025; originally announced October 2025.

    Comments: 39th Conference on Neural Information Processing Systems (NeurIPS 2025) Workshop: AI and ML for Next-Generation Wireless Communications and Networking (AI4NextG)

  28. arXiv:2510.08722  [pdf, ps, other

    cs.LG cs.AI

    Enhancing Self-Supervised Learning with Semantic Pairs A New Dataset and Empirical Study

    Authors: Mohammad Alkhalefi, Georgios Leontidis, Mingjun Zhong

    Abstract: Instance discrimination is a self-supervised representation learning paradigm wherein individual instances within a dataset are treated as distinct classes. This is typically achieved by generating two disparate views of each instance by applying stochastic transformations, encouraging the model to learn representations invariant to the common underlying object across these views. While this appro… ▽ More

    Submitted 13 October, 2025; v1 submitted 9 October, 2025; originally announced October 2025.

    Comments: 16 pages, 7 figures, 5 tables

  29. arXiv:2510.07315  [pdf, ps, other

    cs.CL cs.AI cs.LG cs.SE

    Vibe Checker: Aligning Code Evaluation with Human Preference

    Authors: Ming Zhong, Xiang Zhou, Ting-Yun Chang, Qingze Wang, Nan Xu, Xiance Si, Dan Garrette, Shyam Upadhyay, Jeremiah Liu, Jiawei Han, Benoit Schillings, Jiao Sun

    Abstract: Large Language Models (LLMs) have catalyzed vibe coding, where users leverage LLMs to generate and iteratively refine code through natural language interactions until it passes their vibe check. Vibe check is tied to real-world human preference and goes beyond functionality: the solution should feel right, read cleanly, preserve intent, and remain correct. However, current code evaluation remains… ▽ More

    Submitted 8 October, 2025; originally announced October 2025.

    Comments: Preprint

  30. arXiv:2510.05691  [pdf, ps, other

    cs.CL

    DecEx-RAG: Boosting Agentic Retrieval-Augmented Generation with Decision and Execution Optimization via Process Supervision

    Authors: Yongqi Leng, Yikun Lei, Xikai Liu, Meizhi Zhong, Bojian Xiong, Yurong Zhang, Yan Gao, Yi Wu, Yao Hu, Deyi Xiong

    Abstract: Agentic Retrieval-Augmented Generation (Agentic RAG) enhances the processing capability for complex tasks through dynamic retrieval and adaptive workflows. Recent advances (e.g., Search-R1) have shown that outcome-supervised reinforcement learning demonstrate strong performance. However, this approach still suffers from inefficient exploration, sparse reward signals, and ambiguous global reward fe… ▽ More

    Submitted 7 October, 2025; originally announced October 2025.

  31. arXiv:2510.02360  [pdf, ps, other

    cs.CL cs.AI

    Spiral of Silence in Large Language Model Agents

    Authors: Mingze Zhong, Meng Fang, Zijing Shi, Yuxuan Huang, Shunfeng Zheng, Yali Du, Ling Chen, Jun Wang

    Abstract: The Spiral of Silence (SoS) theory holds that individuals with minority views often refrain from speaking out for fear of social isolation, enabling majority positions to dominate public discourse. When the 'agents' are large language models (LLMs), however, the classical psychological explanation is not directly applicable, since SoS was developed for human societies. This raises a central questi… ▽ More

    Submitted 7 October, 2025; v1 submitted 28 September, 2025; originally announced October 2025.

    Comments: Accepted to EMNLP 2025 (Findings)

  32. arXiv:2510.00387  [pdf

    cs.LG cs.HC

    Bayesian Distributional Models of Executive Functioning

    Authors: Robert Kasumba, Zeyu Lu, Dom CP Marticorena, Mingyang Zhong, Paul Beggs, Anja Pahor, Geetha Ramani, Imani Goffney, Susanne M Jaeggi, Aaron R Seitz, Jacob R Gardner, Dennis L Barbour

    Abstract: This study uses controlled simulations with known ground-truth parameters to evaluate how Distributional Latent Variable Models (DLVM) and Bayesian Distributional Active LEarning (DALE) perform in comparison to conventional Independent Maximum Likelihood Estimation (IMLE). DLVM integrates observations across multiple executive function tasks and individuals, allowing parameter estimation even unde… ▽ More

    Submitted 7 October, 2025; v1 submitted 30 September, 2025; originally announced October 2025.

    Comments: 42 pages, 8 figures, 1 table

  33. arXiv:2509.23716  [pdf, ps, other

    cs.SI physics.soc-ph

    Robustness of One-to-Many Interdependent Higher-order Networks Against Cascading Failures

    Authors: Cheng Qian, Dandan Zhao, Bo Zhang, Ming Zhong, Jianmin Han, Shenghong Li, Hao Peng, Wei Wang

    Abstract: In the real world, the stable operation of a network is usually inseparable from the mutual support of other networks. In such an interdependent network, a node in one layer may depend on multiple nodes in another layer, forming a complex one-to-many dependency relationship. Meanwhile, there may also be higher-order interactions between multiple nodes within a layer, which increases the connectivi… ▽ More

    Submitted 28 September, 2025; originally announced September 2025.

  34. arXiv:2509.17946  [pdf, ps, other

    cs.CL cs.AI cs.HC

    HICode: Hierarchical Inductive Coding with LLMs

    Authors: Mian Zhong, Pristina Wang, Anjalie Field

    Abstract: Despite numerous applications for fine-grained corpus analysis, researchers continue to rely on manual labeling, which does not scale, or statistical tools like topic modeling, which are difficult to control. We propose that LLMs have the potential to scale the nuanced analyses that researchers typically conduct manually to large text corpora. To this effect, inspired by qualitative research metho… ▽ More

    Submitted 22 September, 2025; originally announced September 2025.

    Comments: Long paper accepted at EMNLP 2025 main conference, 19 pages, 8 figures

  35. A Survey on Retrieval And Structuring Augmented Generation with Large Language Models

    Authors: Pengcheng Jiang, Siru Ouyang, Yizhu Jiao, Ming Zhong, Runchu Tian, Jiawei Han

    Abstract: Large Language Models (LLMs) have revolutionized natural language processing with their remarkable capabilities in text generation and reasoning. However, these models face critical challenges when deployed in real-world applications, including hallucination generation, outdated knowledge, and limited domain expertise. Retrieval And Structuring (RAS) Augmented Generation addresses these limitation… ▽ More

    Submitted 12 September, 2025; originally announced September 2025.

    Comments: KDD'25 survey track

  36. arXiv:2509.00425  [pdf, ps, other

    cs.CL

    The Gold Medals in an Empty Room: Diagnosing Metalinguistic Reasoning in LLMs with Camlang

    Authors: Fenghua Liu, Yulong Chen, Yixuan Liu, Zhujun Jin, Solomon Tsai, Ming Zhong

    Abstract: Large Language Models (LLMs) achieve gold-medal performance across many benchmarks, yet it remains unclear whether such success reflects genuine reasoning or pattern matching. From a cognitive science perspective, an informative test is whether models can master an unfamiliar language through explicit metalinguistic deductive learning, a paradigm where human learners can reliably internalise gramm… ▽ More

    Submitted 30 August, 2025; originally announced September 2025.

    Comments: Working in progress

  37. arXiv:2508.18202  [pdf, ps, other

    physics.flu-dyn cs.MS math.NA physics.comp-ph

    Uncertain data assimilation for urban wind flow simulations with OpenLB-UQ

    Authors: Mingliang Zhong, Dennis Teutscher, Adrian Kummerländer, Mathias J. Krause, Martin Frank, Stephan Simonis

    Abstract: Accurate prediction of urban wind flow is essential for urban planning, pedestrian safety, and environmental management. Yet, it remains challenging due to uncertain boundary conditions and the high cost of conventional CFD simulations. This paper presents the use of the modular and efficient uncertainty quantification (UQ) framework OpenLB-UQ for urban wind flow simulations. We specifically use t… ▽ More

    Submitted 25 August, 2025; originally announced August 2025.

  38. arXiv:2508.15793  [pdf, ps, other

    cs.CL cs.LG

    Format as a Prior: Quantifying and Analyzing Bias in LLMs for Heterogeneous Data

    Authors: Jiacheng Liu, Mayi Xu, Qiankun Pi, Wenli Li, Ming Zhong, Yuanyuan Zhu, Mengchi Liu, Tieyun Qian

    Abstract: Large Language Models (LLMs) are increasingly employed in applications that require processing information from heterogeneous formats, including texts, tables, infoboxes, and knowledge graphs. However, systematic biases toward particular formats may undermine LLMs' ability to integrate heterogeneous data impartially, potentially resulting in reasoning errors and increased risks in downstream tasks… ▽ More

    Submitted 13 January, 2026; v1 submitted 12 August, 2025; originally announced August 2025.

    Comments: Accepted by AAAI 2026, camera ready version

  39. arXiv:2508.13867  [pdf, ps, other

    physics.flu-dyn cs.MS math.NA physics.comp-ph

    OpenLB-UQ: An Uncertainty Quantification Framework for Incompressible Fluid Flow Simulations

    Authors: Mingliang Zhong, Adrian Kummerländer, Shota Ito, Mathias J. Krause, Martin Frank, Stephan Simonis

    Abstract: Uncertainty quantification (UQ) is crucial in computational fluid dynamics to assess the reliability and robustness of simulations, given the uncertainties in input parameters. OpenLB is an open-source lattice Boltzmann method library designed for efficient and extensible simulations of complex fluid dynamics on high-performance computers. In this work, we leverage the efficiency of OpenLB for lar… ▽ More

    Submitted 19 August, 2025; originally announced August 2025.

  40. arXiv:2508.13333  [pdf, ps, other

    cs.AI cs.NE math.OC

    HiFo-Prompt: Prompting with Hindsight and Foresight for LLM-based Automatic Heuristic Design

    Authors: Chentong Chen, Mengyuan Zhong, Ye Fan, Jialong Shi, Jianyong Sun

    Abstract: LLM-based Automatic Heuristic Design (AHD) within Evolutionary Computation (EC) frameworks has shown promising results. However, its effectiveness is hindered by the use of static operators and the lack of knowledge accumulation mechanisms. We introduce HiFo-Prompt, a framework that guides LLMs with two synergistic prompting strategies: Foresight and Hindsight. Foresight-based prompts adaptively s… ▽ More

    Submitted 8 February, 2026; v1 submitted 18 August, 2025; originally announced August 2025.

    Comments: Accepted at ICLR 2026

  41. arXiv:2508.09473  [pdf, ps, other

    cs.LG cs.AI cs.CL

    NeuronTune: Fine-Grained Neuron Modulation for Balanced Safety-Utility Alignment in LLMs

    Authors: Birong Pan, Mayi Xu, Qiankun Pi, Jianhao Chen, Yuanyuan Zhu, Ming Zhong, Tieyun Qian

    Abstract: Ensuring robust safety alignment while preserving utility is critical for the reliable deployment of Large Language Models (LLMs). However, current techniques fundamentally suffer from intertwined deficiencies: insufficient robustness against malicious attacks, frequent refusal of benign queries, degradation in generated text quality and general task performance--the former two reflecting deficits… ▽ More

    Submitted 13 August, 2025; originally announced August 2025.

  42. arXiv:2508.09016  [pdf, ps, other

    cs.CL cs.LG

    A Survey on Training-free Alignment of Large Language Models

    Authors: Birong Pan, Yongqi Li, Weiyu Zhang, Wenpeng Lu, Mayi Xu, Shen Zhou, Yuanyuan Zhu, Ming Zhong, Tieyun Qian

    Abstract: The alignment of large language models (LLMs) aims to ensure their outputs adhere to human values, ethical standards, and legal norms. Traditional alignment methods often rely on resource-intensive fine-tuning (FT), which may suffer from knowledge degradation and face challenges in scenarios where the model accessibility or computational resources are constrained. In contrast, training-free (TF) a… ▽ More

    Submitted 10 September, 2025; v1 submitted 12 August, 2025; originally announced August 2025.

    Comments: Accepted to EMNLP 2025 (findings), camera-ready version

  43. arXiv:2508.08785  [pdf, ps, other

    cs.CL

    Privacy-protected Retrieval-Augmented Generation for Knowledge Graph Question Answering

    Authors: Yunfeng Ning, Mayi Xu, Jintao Wen, Qiankun Pi, Yuanyuan Zhu, Ming Zhong, Jiawei Jiang, Tieyun Qian

    Abstract: LLMs often suffer from hallucinations and outdated or incomplete knowledge. RAG is proposed to address these issues by integrating external knowledge like that in KGs into LLMs. However, leveraging private KGs in RAG systems poses significant privacy risks due to the black-box nature of LLMs and potential insecure data transmission, especially when using third-party LLM APIs lacking transparency a… ▽ More

    Submitted 3 December, 2025; v1 submitted 12 August, 2025; originally announced August 2025.

    Comments: Accepted by AAAI 2026, camera ready version

  44. Caption: Generating Informative Content Labels for Image Buttons Using Next-Screen Context

    Authors: Mingyuan Zhong, Ajit Mallavarapu, Qing Nie

    Abstract: We present Caption, an LLM-powered content label generation tool for visual interactive elements on mobile devices. Content labels are essential for screen readers to provide announcements for image-based elements, but are often missing or uninformative due to developer neglect. Automated captioning systems attempt to address this, but are limited to on-screen context, often resulting in inaccurat… ▽ More

    Submitted 12 August, 2025; originally announced August 2025.

  45. SlideAudit: A Dataset and Taxonomy for Automated Evaluation of Presentation Slides

    Authors: Zhuohao Jerry Zhang, Ruiqi Chen, Mingyuan Zhong, Jacob O. Wobbrock

    Abstract: Automated evaluation of specific graphic designs like presentation slides is an open problem. We present SlideAudit, a dataset for automated slide evaluation. We collaborated with design experts to develop a thorough taxonomy of slide design flaws. Our dataset comprises 2400 slides collected and synthesized from multiple sources, including a subset intentionally modified with specific design probl… ▽ More

    Submitted 5 August, 2025; originally announced August 2025.

    Comments: UIST 2025

  46. arXiv:2507.23118  [pdf, ps, other

    cs.SE

    FlowETL: An Autonomous Example-Driven Pipeline for Data Engineering

    Authors: Mattia Di Profio, Mingjun Zhong, Yaji Sripada, Marcel Jaspars

    Abstract: The Extract, Transform, Load (ETL) workflow is fundamental for populating and maintaining data warehouses and other data stores accessed by analysts for downstream tasks. A major shortcoming of modern ETL solutions is the extensive need for a human-in-the-loop, required to design and implement context-specific, and often non-generalisable transformations. While related work in the field of ETL aut… ▽ More

    Submitted 30 July, 2025; originally announced July 2025.

  47. arXiv:2507.15715  [pdf, ps, other

    cs.CL astro-ph.IM

    From Queries to Criteria: Understanding How Astronomers Evaluate LLMs

    Authors: Alina Hyk, Kiera McCormick, Mian Zhong, Ioana Ciucă, Sanjib Sharma, John F Wu, J. E. G. Peek, Kartheik G. Iyer, Ziang Xiao, Anjalie Field

    Abstract: There is growing interest in leveraging LLMs to aid in astronomy and other scientific research, but benchmarks for LLM evaluation in general have not kept pace with the increasingly diverse ways that real people evaluate and use these models. In this study, we seek to improve evaluation procedures by building an understanding of how users evaluate LLMs. We focus on a particular use case: an LLM-po… ▽ More

    Submitted 5 August, 2025; v1 submitted 21 July, 2025; originally announced July 2025.

    Comments: Accepted to the Conference on Language Modeling 2025 (COLM), 22 pages, 6 figures

  48. arXiv:2506.13992  [pdf, ps, other

    cs.LG cs.AI cs.CL stat.ME

    AssistedDS: Benchmarking How External Domain Knowledge Assists LLMs in Automated Data Science

    Authors: An Luo, Xun Xian, Jin Du, Fangqiao Tian, Ganghua Wang, Ming Zhong, Shengchun Zhao, Xuan Bi, Zirui Liu, Jiawei Zhou, Jayanth Srinivasa, Ashish Kundu, Charles Fleming, Mingyi Hong, Jie Ding

    Abstract: Large language models (LLMs) have advanced the automation of data science workflows. Yet it remains unclear whether they can critically leverage external domain knowledge as human data scientists do in practice. To answer this question, we introduce AssistedDS (Assisted Data Science), a benchmark designed to systematically evaluate how LLMs handle domain knowledge in tabular prediction tasks. Assi… ▽ More

    Submitted 22 October, 2025; v1 submitted 25 May, 2025; originally announced June 2025.

    MSC Class: 62-07; 62-08; 68T05; 68T07; 68T01; 68T50 ACM Class: I.2.0; I.2.6; I.2.7; I.5.1; I.5.4; H.2.8; G.3

  49. arXiv:2506.00930  [pdf, ps, other

    cs.AI cs.CL

    Aligning VLM Assistants with Personalized Situated Cognition

    Authors: Yongqi Li, Shen Zhou, Xiaohu Li, Xin Miao, Jintao Wen, Mayi Xu, Jianhao Chen, Birong Pan, Hankun Kang, Yuanyuan Zhu, Ming Zhong, Tieyun Qian

    Abstract: Vision-language models (VLMs) aligned with general human objectives, such as being harmless and hallucination-free, have become valuable assistants of humans in managing visual tasks. However, people with diversified backgrounds have different cognition even in the same situation. Consequently, they may have personalized expectations for VLM assistants. This highlights the urgent need to align VLM… ▽ More

    Submitted 1 June, 2025; originally announced June 2025.

    Comments: Accepted to ACL 2025 (main), camera-ready version

  50. arXiv:2505.14180  [pdf

    cs.IR cs.CV

    Bridge the Gap between Past and Future: Siamese Model Optimization for Context-Aware Document Ranking

    Authors: Songhao Wu, Quan Tu, Mingjie Zhong, Hong Liu, Jia Xu, Jinjie Gu, Rui Yan

    Abstract: In the realm of information retrieval, users often engage in multi-turn interactions with search engines to acquire information, leading to the formation of sequences of user feedback behaviors. Leveraging the session context has proven to be beneficial for inferring user search intent and document ranking. A multitude of approaches have been proposed to exploit in-session context for improved doc… ▽ More

    Submitted 20 May, 2025; originally announced May 2025.

    ACM Class: H.3.3