Skip to main content

🏢 Tsinghua University

Can LLMs Learn by Teaching for Better Reasoning? A Preliminary Study
·5164 words·25 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Tsinghua University
LLMs can improve reasoning by teaching weaker models, a process called Learning by Teaching (LbT), as shown in this preliminary study. LbT enhances not just student models, but also the teacher model…
C-GAIL: Stabilizing Generative Adversarial Imitation Learning with Control Theory
·1787 words·9 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Tsinghua University
C-GAIL stabilizes Generative Adversarial Imitation Learning by applying control theory, resulting in faster convergence, reduced oscillation, and better expert policy matching.
Bridging the Divide: Reconsidering Softmax and Linear Attention
·2335 words·11 mins· loading · loading
Computer Vision Image Classification 🏢 Tsinghua University
InLine attention, a novel method, bridges the performance gap between softmax and linear attention by incorporating injectivity and local modeling, achieving superior performance while maintaining lin…
Bridging Multicalibration and Out-of-distribution Generalization Beyond Covariate Shift
·1648 words·8 mins· loading · loading
AI Theory Generalization 🏢 Tsinghua University
New model-agnostic framework for out-of-distribution generalization uses multicalibration across overlapping groups, showing improved robustness and prediction under various distribution shifts.
Boosting the Potential of Large Language Models with an Intelligent Information Assistant
·1837 words·9 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Tsinghua University
Boosting LLMs with an intelligent information assistant, ASSISTRAG, significantly improves accuracy and reasoning, especially for less advanced models.
BoostAdapter: Improving Vision-Language Test-Time Adaptation via Regional Bootstrapping
·3371 words·16 mins· loading · loading
AI Generated Multimodal Learning Vision-Language Models 🏢 Tsinghua University
BoostAdapter enhances vision-language model test-time adaptation by combining instance-agnostic historical samples with instance-aware boosting samples for superior out-of-distribution and cross-domai…
Binocular-Guided 3D Gaussian Splatting with View Consistency for Sparse View Synthesis
·2801 words·14 mins· loading · loading
Computer Vision 3D Vision 🏢 Tsinghua University
Binocular-guided 3D Gaussian splatting with self-supervision generates high-quality novel views from sparse inputs without external priors, significantly outperforming state-of-the-art methods.
AutoTimes: Autoregressive Time Series Forecasters via Large Language Models
·5046 words·24 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Tsinghua University
AutoTimes repurposes LLMs as autoregressive time series forecasters, achieving state-of-the-art results with minimal trainable parameters and faster training/inference.
Autonomous Agents for Collaborative Task under Information Asymmetry
·3171 words·15 mins· loading · loading
AI Generated Natural Language Processing Dialogue Systems 🏢 Tsinghua University
iAgents: a novel multi-agent system leveraging LLMs, overcomes information asymmetry by mirroring human social networks to enable effective collaboration in complex tasks, achieving high accuracy in d…
Assembly Fuzzy Representation on Hypergraph for Open-Set 3D Object Retrieval
·2034 words·10 mins· loading · loading
Computer Vision 3D Vision 🏢 Tsinghua University
Hypergraph-Based Assembly Fuzzy Representation (HAFR) excels at open-set 3D object retrieval by using part-level shapes and fuzzy representations to overcome challenges posed by unseen object categori…
An Efficient Memory Module for Graph Few-Shot Class-Incremental Learning
·2278 words·11 mins· loading · loading
Machine Learning Few-Shot Learning 🏢 Tsinghua University
Mecoin: a novel memory module for efficient graph few-shot class-incremental learning, tackles catastrophic forgetting by employing structured memory units and a memory representation adaptation modul…
AMOR: A Recipe for Building Adaptable Modular Knowledge Agents Through Process Feedback
·3399 words·16 mins· loading · loading
Natural Language Processing Question Answering 🏢 Tsinghua University
AMOR: Adaptable Modular knowledge agent using LLMs, excels with FSM-based reasoning and process feedback, enabling human supervision and domain adaptation.
AlphaTablets: A Generic Plane Representation for 3D Planar Reconstruction from Monocular Videos
·2409 words·12 mins· loading · loading
AI Generated Computer Vision 3D Vision 🏢 Tsinghua University
AlphaTablets revolutionizes 3D planar reconstruction from monocular videos with its novel rectangle-based representation featuring continuous surfaces and precise boundaries, achieving state-of-the-ar…
Aligning Diffusion Behaviors with Q-functions for Efficient Continuous Control
·1969 words·10 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Tsinghua University
Efficient Diffusion Alignment (EDA) leverages pretrained diffusion models and Q-functions for efficient continuous control, exceeding all baselines with minimal annotation.
AdaPKC: PeakConv with Adaptive Peak Receptive Field for Radar Semantic Segmentation
·2841 words·14 mins· loading · loading
Computer Vision Image Segmentation 🏢 Tsinghua University
AdaPKC upgrades PeakConv for superior radar semantic segmentation by dynamically adjusting its receptive field, outperforming current state-of-the-art methods.
3D Structure Prediction of Atomic Systems with Flow-based Direct Preference Optimization
·2483 words·12 mins· loading · loading
AI Generated AI Applications Healthcare 🏢 Tsinghua University
FlowDPO: Revolutionizing 3D structure prediction with flexible probability paths & Direct Preference Optimization for enhanced accuracy and reduced hallucinations.