Skip to main content

Posters

2024

Mitigating Covariate Shift in Behavioral Cloning via Robust Stationary Distribution Correction
·2650 words·13 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 KAIST
DrilDICE robustly tackles covariate shift in offline imitation learning by using a stationary distribution correction and a distributionally robust objective, significantly improving performance.
Mitigating Biases in Blackbox Feature Extractors for Image Classification Tasks
·2583 words·13 mins· loading · loading
Computer Vision Image Classification 🏢 Indian Institute of Science
Researchers propose a simple yet effective clustering-based adaptive margin loss to mitigate biases inherited by black-box feature extractors in image classification tasks.
Mitigating Backdoor Attack by Injecting Proactive Defensive Backdoor
·3713 words·18 mins· loading · loading
Machine Learning Deep Learning 🏢 School of Data Science
Proactive Defensive Backdoor (PDB) thwarts malicious backdoors by injecting a hidden defensive backdoor during training, suppressing attacks while maintaining model utility.
MiSO: Optimizing brain stimulation to create neural activity states
·2684 words·13 mins· loading · loading
AI Generated Machine Learning Deep Learning 🏢 Carnegie Mellon University
MiSO: a novel closed-loop brain stimulation framework optimizes stimulation parameters to achieve desired neural population activity states, overcoming limitations of current methods by merging data a…
Mining and Transferring Feature-Geometry Coherence for Unsupervised Point Cloud Registration
·2514 words·12 mins· loading · loading
AI Generated Computer Vision 3D Vision 🏢 Fujian Key Laboratory of Sensing and Computing for Smart Cities, Xiamen University, China
INTEGER: a novel unsupervised point cloud registration method leveraging feature-geometry coherence for reliable pseudo-label mining and density-invariant feature learning, achieving state-of-the-art …
Minimizing UCB: a Better Local Search Strategy in Local Bayesian Optimization
·1728 words·9 mins· loading · loading
Machine Learning Optimization 🏢 Academy of Mathematics and Systems Science, Chinese Academy of Sciences
MinUCB and LA-MinUCB, novel local Bayesian optimization algorithms, replace gradient descent with UCB minimization for efficient, theoretically-sound local search.
Minimax Optimal and Computationally Efficient Algorithms for Distributionally Robust Offline Reinforcement Learning
·1835 words·9 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏢 Duke University
Minimax-optimal, computationally efficient algorithms are proposed for distributionally robust offline reinforcement learning, addressing challenges posed by function approximation and model uncertain…
MiniCache: KV Cache Compression in Depth Dimension for Large Language Models
·3037 words·15 mins· loading · loading
Natural Language Processing Large Language Models 🏢 ZIP Lab, Monash University
MiniCache: A novel approach to drastically reduce LLM KV cache memory footprint.
Mini-Sequence Transformers: Optimizing Intermediate Memory for Long Sequences Training
·2712 words·13 mins· loading · loading
Natural Language Processing Large Language Models 🏢 California Institute of Technology
MINI-SEQUENCE TRANSFORMER (MST) drastically reduces memory usage in LLM training by processing mini-sequences iteratively, enabling training with 12-24x longer sequences than conventional methods with…
MindMerger: Efficiently Boosting LLM Reasoning in non-English Languages
·2639 words·13 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Shanghai Artificial Intelligence Laboratory
MindMerger efficiently boosts LLM reasoning in non-English languages by merging LLMs with external multilingual language understanding capabilities, achieving significant accuracy improvements, especi…
Mind's Eye of LLMs: Visualization-of-Thought Elicits Spatial Reasoning in Large Language Models
·4302 words·21 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Microsoft Research
LLMs’ spatial reasoning abilities are boosted by visualizing their thought processes via ‘Visualization-of-Thought’ prompting, significantly improving performance on navigation and tiling tasks.
Mind the Graph When Balancing Data for Fairness or Robustness
·1841 words·9 mins· loading · loading
AI Theory Fairness 🏢 Google DeepMind
Data balancing in machine learning can hurt fairness and robustness; this paper reveals when and why, offering solutions for safer AI.
Mind the Gap: A Causal Perspective on Bias Amplification in Prediction & Decision-Making
·1773 words·9 mins· loading · loading
AI Theory Fairness 🏢 Columbia University
AI bias amplification in decision-making is uncovered, showing how fair prediction scores can become discriminatory after thresholding, urging stronger regulatory oversight.
Mind the Gap Between Prototypes and Images in Cross-domain Finetuning
·3462 words·17 mins· loading · loading
Computer Vision Few-Shot Learning 🏢 Hong Kong Baptist University
CoPA improves cross-domain few-shot learning by adapting separate transformations for prototype and image embeddings, significantly enhancing performance and revealing better representation clusters.
MimicTalk: Mimicking a personalized and expressive 3D talking face in minutes
·1847 words·9 mins· loading · loading
Computer Vision Image Generation 🏢 Zhejiang University
MimicTalk generates realistic, expressive talking videos in minutes using a pre-trained model adapted to individual identities.
Mimicking To Dominate: Imitation Learning Strategies for Success in Multiagent Games
·1956 words·10 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏢 Singapore Management University
IMAX-PPO: A novel multi-agent RL algorithm leveraging imitation learning to predict opponent actions, achieving superior performance in complex games.
MILP-StuDio: MILP Instance Generation via Block Structure Decomposition
·3596 words·17 mins· loading · loading
AI Theory Optimization 🏢 University of Science and Technology of China
MILP-StuDio generates high-quality mixed-integer linear programming instances by preserving crucial block structures, significantly improving learning-based solver performance.
MIDGArD: Modular Interpretable Diffusion over Graphs for Articulated Designs
·2491 words·12 mins· loading · loading
Computer Vision 3D Vision 🏢 Intel Labs
MIDGARD: Generate high-quality, simulatable 3D articulated assets with enhanced control and interpretability using a novel diffusion-based framework.
Microstructures and Accuracy of Graph Recall by Large Language Models
·2595 words·13 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Cornell University
LLMs struggle with graph recall, exhibiting biases like favoring triangles and underperforming compared to humans; advanced models show striking domain dependence.
MicroAdam: Accurate Adaptive Optimization with Low Space Overhead and Provable Convergence
·2050 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Institute of Science and Technology Austria (ISTA)
MICROADAM: A new Adam optimizer variant dramatically cuts memory usage for training large language models without compromising accuracy or provable convergence.