🏢 Peking University
Efficient Multi-task LLM Quantization and Serving for Multiple LoRA Adapters
·2138 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Peking University
LoRA-Inlaid: a novel multi-task LLM serving system boosts throughput by 1.58x, latency by 1.76x, and job completion time by 2x, while improving SLO attainment by 10x, all while maintaining model quali…
Du-IN: Discrete units-guided mask modeling for decoding speech from Intracranial Neural signals
·3577 words·17 mins·
loading
·
loading
AI Applications
Healthcare
🏢 Peking University
Du-IN: Revolutionizing speech decoding from intracranial neural signals with discrete units-guided mask modeling, achieving state-of-the-art performance on a challenging Chinese word-reading sEEG dat…
Distribution Guidance Network for Weakly Supervised Point Cloud Semantic Segmentation
·2253 words·11 mins·
loading
·
loading
Computer Vision
3D Vision
🏢 Peking University
DGNet enhances weakly supervised point cloud segmentation by aligning feature embeddings to a mixture of von Mises-Fisher distributions, achieving state-of-the-art performance.
Dissecting the Failure of Invariant Learning on Graphs
·4452 words·21 mins·
loading
·
loading
AI Generated
AI Theory
Generalization
🏢 Peking University
Cross-environment Intra-class Alignment (CIA) and its label-free variant, CIA-LRA, significantly improve node-level OOD generalization on graphs by aligning representations and eliminating spurious fe…
DeltaDock: A Unified Framework for Accurate, Efficient, and Physically Reliable Molecular Docking
·4235 words·20 mins·
loading
·
loading
AI Generated
AI Applications
Healthcare
🏢 Peking University
DeltaDock: a novel two-stage framework revolutionizes molecular docking by improving accuracy and reliability, achieving a 300% increase in success rate compared to the prior state-of-the-art in blind…
Delta-CoMe: Training-Free Delta-Compression with Mixed-Precision for Large Language Models
·2535 words·12 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Peking University
Delta-CoMe: Training-free mixed-precision delta compression boosts LLM deployment efficiency.
Continuous Spatiotemporal Events Decoupling through Spike-based Bayesian Computation
·1859 words·9 mins·
loading
·
loading
Computer Vision
Image Segmentation
🏢 Peking University
Spiking neural network effectively segments mixed-motion event streams via spike-based Bayesian computation, achieving efficient real-time motion decoupling.
Contextual Decision-Making with Knapsacks Beyond the Worst Case
·450 words·3 mins·
loading
·
loading
AI Theory
Optimization
🏢 Peking University
This work unveils a novel algorithm for contextual decision-making with knapsacks, achieving significantly improved regret bounds beyond worst-case scenarios, thereby offering a more practical and eff…
Categorical Flow Matching on Statistical Manifolds
·2341 words·11 mins·
loading
·
loading
AI Generated
Machine Learning
Generative Models
🏢 Peking University
Statistical Flow Matching (SFM) uses information geometry to create a new flow-matching framework for generating discrete data, achieving superior sampling quality and likelihood compared to existing …
Can Graph Learning Improve Planning in LLM-based Agents?
·2929 words·14 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Peking University
GNNs enhance LLM-based task planning by improving the ability to process task graphs, surpassing existing solutions even without training.
Buffer of Thoughts: Thought-Augmented Reasoning with Large Language Models
·2186 words·11 mins·
loading
·
loading
Large Language Models
🏢 Peking University
Buffer of Thoughts (BoT) boosts Large Language Model reasoning by storing and reusing high-level ’thought-templates’, achieving significant accuracy and efficiency gains across diverse tasks.
Bridging Geometric States via Geometric Diffusion Bridge
·1526 words·8 mins·
loading
·
loading
Machine Learning
Deep Learning
🏢 Peking University
Geometric Diffusion Bridge (GDB) accurately predicts geometric state evolution in complex systems by leveraging a probabilistic approach and equivariant diffusion processes, surpassing existing deep l…
AutoSurvey: Large Language Models Can Automatically Write Surveys
·2587 words·13 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Peking University
AutoSurvey automates comprehensive literature survey creation using LLMs, overcoming challenges of context limitations and knowledge constraints via a novel, efficient, and rigorously evaluated method…
Autoformalize Mathematical Statements by Symbolic Equivalence and Semantic Consistency
·2235 words·11 mins·
loading
·
loading
AI Generated
Natural Language Processing
Large Language Models
🏢 Peking University
Boosting AI’s math skills, this paper introduces a novel framework for autoformalizing mathematical statements, improving accuracy by 0.22-1.35x via symbolic equivalence and semantic consistency check…
ArkVale: Efficient Generative LLM Inference with Recallable Key-Value Eviction
·2152 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Peking University
ARKVALE boosts LLM inference efficiency by intelligently evicting and recalling key-value pairs from cache, improving latency and throughput without significant accuracy loss.
An Expectation-Maximization Algorithm for Training Clean Diffusion Models from Corrupted Observations
·3657 words·18 mins·
loading
·
loading
AI Generated
Computer Vision
Image Generation
🏢 Peking University
EMDiffusion trains clean diffusion models from corrupted data using an expectation-maximization algorithm, achieving state-of-the-art results on diverse imaging tasks.
Aligner: Efficient Alignment by Learning to Correct
·3091 words·15 mins·
loading
·
loading
Large Language Models
🏢 Peking University
Aligner efficiently aligns LLMs by learning to correct initial responses, achieving significant improvements in helpfulness and harmlessness across various models with resource efficiency.
Adversarial Representation Engineering: A General Model Editing Framework for Large Language Models
·1740 words·9 mins·
loading
·
loading
Natural Language Processing
Large Language Models
🏢 Peking University
Adversarial Representation Engineering (ARE) offers a unified, interpretable approach for editing large language models (LLMs) by using a representation sensor as an editing oracle, enhancing model sa…
Addressing Hidden Confounding with Heterogeneous Observational Datasets for Recommendation
·2627 words·13 mins·
loading
·
loading
AI Generated
Machine Learning
Meta Learning
🏢 Peking University
MetaDebias tackles hidden confounding in recommender systems using heterogeneous observational data, achieving state-of-the-art performance without expensive RCT data.
A Tractable Inference Perspective of Offline RL
·2824 words·14 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
🏢 Peking University
Trifle: Tractable inference for Offline RL achieves state-of-the-art results by using tractable generative models to overcome the inference-time suboptimality of existing sequence modeling approaches.