Skip to main content

Posters

2024

Model Based Inference of Synaptic Plasticity Rules
·2178 words·11 mins· loading · loading
Machine Learning Meta Learning 🏢 Janelia Research Campus
New computational method infers complex brain learning rules from experimental data, revealing active forgetting in reward learning.
Mobility-LLM: Learning Visiting Intentions and Travel Preference from Human Mobility Data with Large Language Models
·2805 words·14 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 Beijing Jiaotong University
Mobility-LLM leverages LLMs to analyze human mobility data from check-in sequences, significantly outperforming existing models in location prediction, user identification, and time prediction tasks.
Mobile-Agent-v2: Mobile Device Operation Assistant with Effective Navigation via Multi-Agent Collaboration
·3087 words·15 mins· loading · loading
AI Generated Multimodal Learning Vision-Language Models 🏢 Beijing Jiaotong University
Mobile-Agent-v2 uses a three-agent collaborative framework (planning, decision, reflection) to improve mobile device operation accuracy by over 30%, overcoming the limitations of single-agent architec…
MO-DDN: A Coarse-to-Fine Attribute-based Exploration Agent for Multi-Object Demand-driven Navigation
·4206 words·20 mins· loading · loading
AI Generated Multimodal Learning Embodied AI 🏢 Peking University
MO-DDN: A new benchmark and coarse-to-fine exploration agent boosts embodied AI’s ability to handle multi-object, preference-based task planning.
MMSite: A Multi-modal Framework for the Identification of Active Sites in Proteins
·2970 words·14 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 School of Computer Science, National Engineering Research Center for Multimedia Software and Institute of Artificial Intelligence, Wuhan University
MMSite: a novel multi-modal framework accurately identifies protein active sites using protein sequences and textual descriptions, achieving state-of-the-art performance.
Mixtures of Experts for Audio-Visual Learning
·2112 words·10 mins· loading · loading
Multimodal Learning Audio-Visual Learning 🏢 Fudan University
AVMoE: a novel parameter-efficient transfer learning approach for audio-visual learning, dynamically allocates expert models (unimodal and cross-modal adapters) based on task demands, achieving superi…
Mixture of Tokens: Continuous MoE through Cross-Example Aggregation
·1989 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 University of Warsaw
Mixture of Tokens (MoT) achieves 3x faster LLM training than dense Transformers and matches state-of-the-art MoE performance via continuous token mixing.
Mixture of Scales: Memory-Efficient Token-Adaptive Binarization for Large Language Models
·1888 words·9 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Seoul National University
BinaryMoS: a novel token-adaptive binarization method that boosts LLM accuracy and efficiency by dynamically merging multiple scaling experts for each token.
Mixture of neural fields for heterogeneous reconstruction in cryo-EM
·4281 words·21 mins· loading · loading
AI Generated Computer Vision 3D Vision 🏢 Stanford University
Hydra: a novel cryo-EM reconstruction method resolves both conformational and compositional heterogeneity ab initio, enabling the analysis of complex, unpurified samples with state-of-the-art accuracy…
Mixture of Link Predictors on Graphs
·3247 words·16 mins· loading · loading
Machine Learning Deep Learning 🏢 Shanghai Jiao Tong University
Link-MoE boosts link prediction accuracy by strategically selecting the best model for each node pair, surpassing single-model approaches.
Mixture of In-Context Experts Enhance LLMs' Long Context Awareness
·2372 words·12 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Gaoling School of Artificial Intelligence, Renmin University of China
MoICE, a novel plug-in, significantly enhances LLMs’ long context awareness by dynamically routing attention using multiple RoPE angles, achieving superior performance with high inference efficiency.
Mixture of Experts Meets Prompt-Based Continual Learning
·1840 words·9 mins· loading · loading
Machine Learning Deep Learning 🏢 VinAI Research
Non-linear Residual Gates (NoRGa) boosts prompt-based continual learning by theoretically framing prefix tuning as adding new experts to a pre-trained Mixture-of-Experts model, achieving state-of-the-…
Mixture of Demonstrations for In-Context Learning
·1953 words·10 mins· loading · loading
Natural Language Processing Large Language Models 🏢 University of Virginia
MoD, a novel Mixture of Demonstrations framework, enhances in-context learning by partitioning demonstration pools and employing expert-wise training, achieving state-of-the-art performance.
Mixture of Adversarial LoRAs: Boosting Robust Generalization in Meta-Tuning
·2777 words·14 mins· loading · loading
Computer Vision Few-Shot Learning 🏢 City University of Hong Kong
Boosting Robust Few-Shot Learning with Adversarial Meta-Tuning!
MixEval: Deriving Wisdom of the Crowd from LLM Benchmark Mixtures
·3103 words·15 mins· loading · loading
AI Generated Natural Language Processing Large Language Models 🏢 National University of Singapore
MixEval revolutionizes LLM benchmarking by blending real-world user queries with existing datasets, creating a cost-effective, unbiased, and dynamic evaluation method.
Mixed Dynamics In Linear Networks: Unifying the Lazy and Active Regimes
·521 words·3 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Courant Institute
A new formula unifies lazy and active neural network training regimes, revealing a mixed regime that combines their strengths for faster convergence and low-rank bias.
Mitigating Spurious Correlations via Disagreement Probability
·2000 words·10 mins· loading · loading
AI Generated AI Theory Fairness 🏢 Seoul National University
DPR, a novel bias mitigation method, robustly improves model performance by leveraging disagreement probability without needing bias labels, achieving state-of-the-art results.
Mitigating Reward Overoptimization via Lightweight Uncertainty Estimation
·1697 words·8 mins· loading · loading
Natural Language Processing Large Language Models 🏢 ByteDance Research
ADVPO, a novel method, tackles reward overoptimization in RLHF via a lightweight uncertainty quantification approach, resulting in enhanced LLM performance and alignment with human values.
Mitigating Partial Observability in Decision Processes via the Lambda Discrepancy
·2495 words·12 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 UC Berkeley
New metric, λ-discrepancy, precisely detects & mitigates partial observability in sequential decision processes, significantly boosting reinforcement learning agent performance.
Mitigating Object Hallucination via Concentric Causal Attention
·2174 words·11 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 Nanyang Technological University
Concentric Causal Attention (CCA) significantly reduces object hallucination in LVLMs by cleverly reorganizing visual tokens to mitigate the impact of long-term decay in Rotary Position Encoding.