Skip to main content

Posters

2024

Reinforcement Learning Policy as Macro Regulator Rather than Macro Placer
·2290 words·11 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 Nanjing University
Reinforcement learning refines existing macro placements, enhancing chip design by improving power, performance, and area (PPA) metrics and integrating the often-overlooked metric of regularity.
Reinforcement Learning Guided Semi-Supervised Learning
·1384 words·7 mins· loading · loading
Machine Learning Semi-Supervised Learning 🏒 School of Computer Science, Carleton University
Reinforcement Learning guides a novel semi-supervised learning method, improving model performance by adaptively balancing labeled and unlabeled data.
Reinforced Cross-Domain Knowledge Distillation on Time Series Data
·2657 words·13 mins· loading · loading
Machine Learning Transfer Learning 🏒 Institute for Infocomm Research, A*STAR, Singapore
Reinforced Cross-Domain Knowledge Distillation (RCD-KD) dynamically selects target samples for efficient knowledge transfer from a complex teacher model to a compact student model, achieving superior …
Reimagining Mutual Information for Enhanced Defense against Data Leakage in Collaborative Inference
·1566 words·8 mins· loading · loading
AI Theory Privacy 🏒 Department of Electrical and Computer Engineering, Duke University
InfoScissors defends collaborative inference from data leakage by cleverly reducing the mutual information between model outputs and sensitive device data, thus ensuring robust privacy without comprom…
Regularizing Hidden States Enables Learning Generalizable Reward Model for LLMs
·3598 words·17 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏒 University of Illinois Urbana-Champaign
Regularizing hidden states improves reward model generalization in RLHF for LLMs, boosting accuracy and mitigating over-optimization.
Regularized Q-Learning
·1497 words·8 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 KAIST
RegQ: A novel regularized Q-learning algorithm ensures convergence with linear function approximation, solving a long-standing instability problem in reinforcement learning.
Regularized Conditional Diffusion Model for Multi-Task Preference Alignment
·2209 words·11 mins· loading · loading
Machine Learning Reinforcement Learning 🏒 Institute of Artificial Intelligence (TeleAI), China Telecom
A novel regularized conditional diffusion model enables effective multi-task preference alignment in sequential decision-making by learning unified preference representations and maximizing mutual inf…
Regularized Adaptive Momentum Dual Averaging with an Efficient Inexact Subproblem Solver for Training Structured Neural Network
·2322 words·11 mins· loading · loading
Machine Learning Deep Learning 🏒 National Taiwan University
RAMDA: a new algorithm ensures efficient training of structured neural networks by achieving optimal structure and outstanding predictive performance.
ReGS: Reference-based Controllable Scene Stylization with Gaussian Splatting
·1952 words·10 mins· loading · loading
Computer Vision 3D Vision 🏒 Johns Hopkins University
ReGS: Real-time reference-based 3D scene stylization using Gaussian Splatting for high-fidelity texture editing and free-view navigation.
Regret Minimization in Stackelberg Games with Side Information
·415 words·2 mins· loading · loading
AI Applications Security 🏒 Carnegie Mellon University
This research shows how to improve Stackelberg game strategies by considering side information, achieving no-regret learning in online settings with stochastic contexts or followers.
Regression under demographic parity constraints via unlabeled post-processing
·1578 words·8 mins· loading · loading
AI Generated AI Theory Fairness 🏒 IRT SystemX, Université Gustave Eiffel
Ensuring fair regression predictions without using sensitive attributes? This paper presents a novel post-processing algorithm, achieving demographic parity with strong theoretical guarantees and comp…
RegExplainer: Generating Explanations for Graph Neural Networks in Regression Tasks
·2208 words·11 mins· loading · loading
AI Theory Interpretability 🏒 New Jersey Institute of Technology
RegExplainer unveils a novel method for interpreting graph neural networks in regression tasks, bridging the explanation gap by addressing distribution shifts and tackling continuously ordered decisio…
Refusal in Language Models Is Mediated by a Single Direction
·4093 words·20 mins· loading · loading
AI Theory Safety 🏒 Independent
LLM refusal is surprisingly mediated by a single, easily manipulated direction in the model’s activation space.
Reflective Multi-Agent Collaboration based on Large Language Models
·2567 words·13 mins· loading · loading
Natural Language Processing Large Language Models 🏒 Gaoling School of Artificial Intelligence, Renmin University of China
COPPER enhances LLM-based multi-agent collaboration via a self-reflection mechanism and counterfactual PPO. It improves reflection quality, alleviates credit assignment issues, and shows strong perfo…
ReFIR: Grounding Large Restoration Models with Retrieval Augmentation
·3091 words·15 mins· loading · loading
Computer Vision Image Generation 🏒 Tsinghua University
ReFIR enhances Large Restoration Models’ accuracy by incorporating retrieved images as external knowledge, mitigating hallucination without retraining.
Referring Human Pose and Mask Estimation In the Wild
·2191 words·11 mins· loading · loading
Multimodal Learning Vision-Language Models 🏒 University of Western Australia
RefHuman: a new dataset and UniPHD model achieve state-of-the-art referring human pose and mask estimation in the wild, using text or positional prompts.
Referencing Where to Focus: Improving Visual Grounding with Referential Query
·2958 words·14 mins· loading · loading
Multimodal Learning Vision-Language Models 🏒 National Key Laboratory of Human-Machine Hybrid Augmented Intelligence
RefFormer boosts visual grounding accuracy by intelligently adapting queries using multi-level image features, effectively guiding the decoder towards the target object.
Reference Trustable Decoding: A Training-Free Augmentation Paradigm for Large Language Models
·2065 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏒 Wuhan University
Reference Trustable Decoding (RTD) revolutionizes large language model adaptation by offering a training-free method, enabling efficient and cost-effective task adaptation without parameter adjustment…
RefDrop: Controllable Consistency in Image or Video Generation via Reference Feature Guidance
·4522 words·22 mins· loading · loading
AI Generated Computer Vision Image Generation 🏒 Georgia Tech
RefDrop: A training-free method enhances image and video generation consistency by directly controlling the influence of reference features on the diffusion process, enabling precise manipulation of c…
ReF-LDM: A Latent Diffusion Model for Reference-based Face Image Restoration
·2816 words·14 mins· loading · loading
Computer Vision Image Generation 🏒 MediaTek
ReF-LDM uses reference images to improve the accuracy of face image restoration, achieving high-quality results faithful to the subject’s true appearance.