Skip to main content

Posters

2024

Belief-State Query Policies for User-Aligned Planning under Partial Observability
·1669 words·8 mins· loading · loading
AI Applications Robotics 🏢 Arizona State University
This paper introduces Belief-State Query (BSQ) constraints for user-aligned planning in partially observable settings, providing algorithms with guaranteed user alignment and computational feasibility…
BehaviorGPT: Smart Agent Simulation for Autonomous Driving with Next-Patch Prediction
·2165 words·11 mins· loading · loading
AI Generated AI Applications Autonomous Vehicles 🏢 City University of Hong Kong
BehaviorGPT, a novel autoregressive Transformer, simulates realistic traffic agent behavior by modeling each time step as ‘current’, achieving top results in the 2024 Waymo Open Sim Agents Challenge.
BECAUSE: Bilinear Causal Representation for Generalizable Offline Model-based Reinforcement Learning
·2825 words·14 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Carnegie Mellon University
BECAUSE: a novel algorithm for generalizable offline model-based reinforcement learning that leverages bilinear causal representation to mitigate objective mismatch caused by confounders in offline da…
Beating Adversarial Low-Rank MDPs with Unknown Transition and Bandit Feedback
·355 words·2 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏢 Google Research
New algorithms conquer adversarial low-rank MDPs, improving regret bounds for unknown transitions and bandit feedback.
Be like a Goldfish, Don't Memorize! Mitigating Memorization in Generative LLMs
·2673 words·13 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 University of Maryland
Goldfish Loss: A novel training method for LLMs dramatically reduces memorization without impacting performance, addressing key safety, privacy, and copyright concerns.
Be Confident in What You Know: Bayesian Parameter Efficient Fine-Tuning of Vision Foundation Models
·3138 words·15 mins· loading · loading
Computer Vision Few-Shot Learning 🏢 Rochester Institute of Technology
Bayesian-PEFT boosts vision model accuracy and confidence in few-shot learning by integrating Bayesian components into PEFT, solving the underconfidence problem.
Bayesian Strategic Classification
·315 words·2 mins· loading · loading
AI Theory Optimization 🏢 Stanford University
Learners can improve accuracy in strategic classification by selectively revealing partial classifier information to agents, strategically guiding agent behavior and maximizing accuracy.
Bayesian Optimization of Functions over Node Subsets in Graphs
·3183 words·15 mins· loading · loading
AI Theory Optimization 🏢 University of Oxford
GraphComBO efficiently optimizes functions defined on node subsets within graphs using Bayesian Optimization. It tackles challenges posed by combinatorial complexity and computationally expensive fun…
Bayesian Nonparametrics Meets Data-Driven Distributionally Robust Optimization
·2079 words·10 mins· loading · loading
AI Generated AI Theory Optimization 🏢 University of Texas at Austin
Boost machine learning model robustness by minimizing a novel data-driven risk criterion that blends Bayesian nonparametrics and smooth ambiguity aversion, ensuring superior out-of-sample performance.
Bayesian Domain Adaptation with Gaussian Mixture Domain-Indexing
·2584 words·13 mins· loading · loading
AI Generated Machine Learning Transfer Learning 🏢 Sun Yat-Sen University
GMDI: a novel Bayesian domain adaptation algorithm significantly improves adaptation by dynamically modeling domain indices using Gaussian Mixture Models, outperforming state-of-the-art methods.
Bayesian Adaptive Calibration and Optimal Design
·1874 words·9 mins· loading · loading
Machine Learning Active Learning 🏢 CSIRO's Data61
BACON: a novel Bayesian adaptive calibration and optimal design algorithm maximizes information gain for data-efficient computer model calibration, significantly outperforming existing methods in synt…
Bayes-optimal learning of an extensive-width neural network from quadratically many samples
·1425 words·7 mins· loading · loading
AI Theory Optimization 🏢 ETH Zurich
This study solves a key challenge in neural network learning, deriving a closed-form expression for the Bayes-optimal test error of extensive-width networks with quadratic activation functions from qu…
Batched Energy-Entropy acquisition for Bayesian Optimization
·4543 words·22 mins· loading · loading
Machine Learning Optimization 🏢 Machine Intelligence
BEEBO: a novel acquisition function for Bayesian Optimization, offering superior explore-exploit balance and handling large batches efficiently, even with noisy data.
Base of RoPE Bounds Context Length
·3176 words·15 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Baichuan Inc.
LLM long-context ability is fundamentally limited by RoPE’s base parameter, which determines an absolute lower bound for achievable context length.
Bandits with Ranking Feedback
·1499 words·8 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Politecnico Di Milano
This paper introduces ‘bandits with ranking feedback,’ a novel bandit variation providing ranked feedback instead of numerical rewards. It proves instance-dependent cases require superlogarithmic reg…
Bandits with Preference Feedback: A Stackelberg Game Perspective
·1514 words·8 mins· loading · loading
Machine Learning Optimization 🏢 ETH Zurich
MAXMINLCB, a novel game-theoretic algorithm, efficiently solves bandit problems with preference feedback over continuous domains, providing anytime-valid, rate-optimal regret guarantees.
Bandits with Abstention under Expert Advice
·2058 words·10 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏢 Alan Turing Institute
The Confidence-Rated Bandits with Abstentions (CBA) algorithm significantly improves reward bounds for prediction with expert advice by strategically leveraging an abstention action.
Bandit-Feedback Online Multiclass Classification: Variants and Tradeoffs
·455 words·3 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Faculty of Computer Science,Technion,Israel
This paper reveals the optimal mistake bounds for online multiclass classification under bandit feedback, showing the cost of limited feedback is at most O(k) times higher than full information, where…
Banded Square Root Matrix Factorization for Differentially Private Model Training
·4880 words·23 mins· loading · loading
AI Theory Privacy 🏢 Institute of Science and Technology (ISTA)
This paper introduces BSR, a novel banded square root matrix factorization for differentially private model training. Unlike existing methods, BSR avoids computationally expensive optimization, enabli…
BAN: Detecting Backdoors Activated by Neuron Noise
·2897 words·14 mins· loading · loading
Machine Learning Deep Learning 🏢 Radboud University
BAN: a novel backdoor defense using adversarial neuron noise for efficient detection and mitigation.