Skip to main content

Showing 1–16 of 16 results for author: Fei, W

Searching in archive cs. Search in all archives.
.
  1. arXiv:2511.14258  [pdf, ps, other

    cs.CL

    Entropy-Guided Reasoning Compression

    Authors: Hourun Zhu, Yang Gao, Wenlong Fei, Jiawei Li, Huashan Sun

    Abstract: Large reasoning models have demonstrated remarkable performance on complex reasoning tasks, yet the excessive length of their chain-of-thought outputs remains a major practical bottleneck due to high computation cost and poor deployability. Existing compression methods have achieved partial success but overlook a crucial phenomenon in the training process -- the entropy conflict. During compressio… ▽ More

    Submitted 24 November, 2025; v1 submitted 18 November, 2025; originally announced November 2025.

    Comments: 10pages, 4 figures

  2. arXiv:2507.18028  [pdf, ps, other

    cs.CL cs.AI

    NeuralDB: Scaling Knowledge Editing in LLMs to 100,000 Facts with Neural KV Database

    Authors: Weizhi Fei, Hao Shi, Jing Xu, Jingchen Peng, Jiazheng Li, Jingzhao Zhang, Bo Bai, Wei Han, Zhenyuan Chen, Xueyan Niu

    Abstract: Efficiently editing knowledge stored in large language models (LLMs) enables model updates without large-scale training. One possible solution is Locate-and-Edit (L\&E), allowing simultaneous modifications of a massive number of facts. However, such editing may compromise the general abilities of LLMs and even result in forgetting edited facts when scaling up to thousands of edits. In this paper,… ▽ More

    Submitted 23 July, 2025; originally announced July 2025.

  3. arXiv:2507.01551  [pdf, ps, other

    cs.LG cs.AI cs.CL

    Self-Guided Process Reward Optimization with Redefined Step-wise Advantage for Process Reinforcement Learning

    Authors: Wu Fei, Hao Kong, Shuxian Liang, Yang Lin, Yibo Yang, Jing Tang, Lei Chen, Xiansheng Hua

    Abstract: Process Reinforcement Learning~(PRL) has demonstrated considerable potential in enhancing the reasoning capabilities of Large Language Models~(LLMs). However, introducing additional process reward models incurs substantial computational overhead, and there is no unified theoretical framework for process-level advantage estimation. To bridge this gap, we propose \textbf{S}elf-Guided \textbf{P}roces… ▽ More

    Submitted 3 July, 2025; v1 submitted 2 July, 2025; originally announced July 2025.

  4. arXiv:2505.08155  [pdf, other

    cs.AI

    Efficient and Scalable Neural Symbolic Search for Knowledge Graph Complex Query Answering

    Authors: Weizhi Fei, Zihao Wang, hang Yin, Shukai Zhao, Wei Zhang, Yangqiu Song

    Abstract: Complex Query Answering (CQA) aims to retrieve answer sets for complex logical formulas from incomplete knowledge graphs, which is a crucial yet challenging task in knowledge graph reasoning. While neuro-symbolic search utilized neural link predictions achieve superior accuracy, they encounter significant complexity bottlenecks: (i) Data complexity typically scales quadratically with the number of… ▽ More

    Submitted 20 May, 2025; v1 submitted 12 May, 2025; originally announced May 2025.

  5. arXiv:2501.14224  [pdf, other

    cs.AI cs.DB cs.LG

    Top Ten Challenges Towards Agentic Neural Graph Databases

    Authors: Jiaxin Bai, Zihao Wang, Yukun Zhou, Hang Yin, Weizhi Fei, Qi Hu, Zheye Deng, Jiayang Cheng, Tianshi Zheng, Hong Ting Tsang, Yisen Gao, Zhongwei Xie, Yufei Li, Lixin Fan, Binhang Yuan, Wei Wang, Lei Chen, Xiaofang Zhou, Yangqiu Song

    Abstract: Graph databases (GDBs) like Neo4j and TigerGraph excel at handling interconnected data but lack advanced inference capabilities. Neural Graph Databases (NGDBs) address this by integrating Graph Neural Networks (GNNs) for predictive analysis and reasoning over incomplete or noisy data. However, NGDBs rely on predefined queries and lack autonomy and adaptability. This paper introduces Agentic Neural… ▽ More

    Submitted 23 January, 2025; originally announced January 2025.

    Comments: 12 Pages

  6. arXiv:2501.12959  [pdf, other

    cs.CL

    Efficient Prompt Compression with Evaluator Heads for Long-Context Transformer Inference

    Authors: Weizhi Fei, Xueyan Niu, Guoqing Xie, Yingqing Liu, Bo Bai, Wei Han

    Abstract: Although applications involving long-context inputs are crucial for the effective utilization of large language models (LLMs), they also result in increased computational costs and reduced performance. To address this challenge, we propose an efficient, training-free prompt compression method that retains key information within compressed prompts. We identify specific attention heads in transforme… ▽ More

    Submitted 5 February, 2025; v1 submitted 22 January, 2025; originally announced January 2025.

  7. arXiv:2406.12331  [pdf, other

    cs.CL cs.AI

    Retrieval Meets Reasoning: Dynamic In-Context Editing for Long-Text Understanding

    Authors: Weizhi Fei, Xueyan Niu, Guoqing Xie, Yanhua Zhang, Bo Bai, Lei Deng, Wei Han

    Abstract: Current Large Language Models (LLMs) face inherent limitations due to their pre-defined context lengths, which impede their capacity for multi-hop reasoning within extensive textual contexts. While existing techniques like Retrieval-Augmented Generation (RAG) have attempted to bridge this gap by sourcing external information, they fall short when direct answers are not readily available. We introd… ▽ More

    Submitted 18 June, 2024; originally announced June 2024.

  8. arXiv:2403.01508  [pdf, other

    cs.AI

    Extending Complex Logical Queries on Uncertain Knowledge Graphs

    Authors: Weizhi Fei, Zihao Wang, Hang Yin, Yang Duan, Yangqiu Song

    Abstract: The study of machine learning-based logical query answering enables reasoning with large-scale and incomplete knowledge graphs. This paper advances this area of research by addressing the uncertainty inherent in knowledge. While the uncertain nature of knowledge is widely recognized in the real world, it does not align seamlessly with the first-order logic that underpins existing studies. To bridg… ▽ More

    Submitted 20 May, 2025; v1 submitted 3 March, 2024; originally announced March 2024.

    Comments: Accepted by the main conference of ACL 2025

  9. arXiv:2402.16121  [pdf, other

    cs.CV cs.AI

    Towards Accurate Post-training Quantization for Reparameterized Models

    Authors: Luoming Zhang, Yefei He, Wen Fei, Zhenyu Lou, Weijia Wu, YangWei Ying, Hong Zhou

    Abstract: Model reparameterization is a widely accepted technique for improving inference speed without compromising performance. However, current Post-training Quantization (PTQ) methods often lead to significant accuracy degradation when applied to reparameterized models. This is primarily caused by channel-specific and sample-specific outliers, which appear only at specific samples and channels and impac… ▽ More

    Submitted 25 February, 2024; originally announced February 2024.

  10. arXiv:2401.11492  [pdf, other

    cs.CV

    Edge-Enabled Real-time Railway Track Segmentation

    Authors: Chen Chenglin, Wang Fei, Yang Min, Qin Yong, Bai Yun

    Abstract: Accurate and rapid railway track segmentation can assist automatic train driving and is a key step in early warning to fixed or moving obstacles on the railway track. However, certain existing algorithms tailored for track segmentation often struggle to meet the requirements of real-time and efficiency on resource-constrained edge devices. Considering this challenge, we propose an edge-enabled rea… ▽ More

    Submitted 21 January, 2024; originally announced January 2024.

  11. arXiv:2312.09571  [pdf, other

    cs.CL cs.IT

    Extending Context Window of Large Language Models via Semantic Compression

    Authors: Weizhi Fei, Xueyan Niu, Pingyi Zhou, Lu Hou, Bo Bai, Lei Deng, Wei Han

    Abstract: Transformer-based Large Language Models (LLMs) often impose limitations on the length of the text input to ensure the generation of fluent and relevant responses. This constraint restricts their applicability in scenarios involving long texts. We propose a novel semantic compression method that enables generalization to texts that are 6-8 times longer, without incurring significant computational c… ▽ More

    Submitted 15 December, 2023; originally announced December 2023.

  12. arXiv:2310.04836  [pdf, other

    cs.AI

    Dual Grained Quantization: Efficient Fine-Grained Quantization for LLM

    Authors: Luoming Zhang, Wen Fei, Weijia Wu, Yefei He, Zhenyu Lou, Hong Zhou

    Abstract: Large Language Models (LLMs) pose significant hardware challenges related to memory requirements and computational ability. There are two mainstream quantization schemes for LLMs: coarse-grained ($\textit{e.g.,}$ channel-wise) quantization and fine-grained ($\textit{e.g.,}$ group-wise) quantization. Fine-grained quantization has smaller quantization loss, consequently achieving superior performanc… ▽ More

    Submitted 7 October, 2023; originally announced October 2023.

    Comments: 15 pages, 2 figures

  13. arXiv:2307.13701  [pdf, other

    cs.AI cs.DB cs.LG cs.LO

    $\text{EFO}_{k}$-CQA: Towards Knowledge Graph Complex Query Answering beyond Set Operation

    Authors: Hang Yin, Zihao Wang, Weizhi Fei, Yangqiu Song

    Abstract: To answer complex queries on knowledge graphs, logical reasoning over incomplete knowledge is required due to the open-world assumption. Learning-based methods are essential because they are capable of generalizing over unobserved knowledge. Therefore, an appropriate dataset is fundamental to both obtaining and evaluating such methods under this paradigm. In this paper, we propose a comprehensive… ▽ More

    Submitted 15 July, 2023; originally announced July 2023.

  14. arXiv:2306.07848  [pdf, other

    cs.CL cs.MM cs.SD eess.AS

    GEmo-CLAP: Gender-Attribute-Enhanced Contrastive Language-Audio Pretraining for Accurate Speech Emotion Recognition

    Authors: Yu Pan, Yanni Hu, Yuguang Yang, Wen Fei, Jixun Yao, Heng Lu, Lei Ma, Jianjun Zhao

    Abstract: Contrastive cross-modality pretraining has recently exhibited impressive success in diverse fields, whereas there is limited research on their merits in speech emotion recognition (SER). In this paper, we propose GEmo-CLAP, a kind of gender-attribute-enhanced contrastive language-audio pretraining (CLAP) method for SER. Specifically, we first construct an effective emotion CLAP (Emo-CLAP) for SER,… ▽ More

    Submitted 4 December, 2023; v1 submitted 13 June, 2023; originally announced June 2023.

    Comments: 5 pages

  15. arXiv:2305.04034  [pdf, other

    cs.AI cs.DB cs.LG

    Wasserstein-Fisher-Rao Embedding: Logical Query Embeddings with Local Comparison and Global Transport

    Authors: Zihao Wang, Weizhi Fei, Hang Yin, Yangqiu Song, Ginny Y. Wong, Simon See

    Abstract: Answering complex queries on knowledge graphs is important but particularly challenging because of the data incompleteness. Query embedding methods address this issue by learning-based models and simulating logical reasoning with set operators. Previous works focus on specific forms of embeddings, but scoring functions between embeddings are underexplored. In contrast to existing scoring functions… ▽ More

    Submitted 6 May, 2023; originally announced May 2023.

    Comments: Findings in ACL 2023. 16 pages, 6 figures, and 8 tables. Our implementation can be found at https://github.com/HKUST-KnowComp/WFRE

  16. arXiv:2010.09278  [pdf, other

    cs.LG cs.AI cs.CV

    MimicNorm: Weight Mean and Last BN Layer Mimic the Dynamic of Batch Normalization

    Authors: Wen Fei, Wenrui Dai, Chenglin Li, Junni Zou, Hongkai Xiong

    Abstract: Substantial experiments have validated the success of Batch Normalization (BN) Layer in benefiting convergence and generalization. However, BN requires extra memory and float-point calculation. Moreover, BN would be inaccurate on micro-batch, as it depends on batch statistics. In this paper, we address these problems by simplifying BN regularization while keeping two fundamental impacts of BN laye… ▽ More

    Submitted 27 September, 2023; v1 submitted 19 October, 2020; originally announced October 2020.

    Comments: This work has been submitted to the IEEE for possible publication