Skip to main content

Posters

2024

Kangaroo: Lossless Self-Speculative Decoding for Accelerating LLMs via Double Early Exiting
·2148 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Huawei Noah's Ark Lab
Kangaroo: Double early exiting boosts LLM speed!
KALM: Knowledgeable Agents by Offline Reinforcement Learning from Large Language Model Rollouts
·3188 words·15 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 National Key Laboratory for Novel Software Technology, Nanjing University, China
KALM: Knowledgeable agents learn complex tasks from LLMs via offline RL using imaginary rollouts, significantly outperforming baselines.
Kaleidoscope: Learnable Masks for Heterogeneous Multi-agent Reinforcement Learning
·2281 words·11 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Hong Kong University of Science and Technology
Kaleidoscope: Learnable Masks for Heterogeneous MARL achieves high sample efficiency and policy diversity by using learnable masks for adaptive partial parameter sharing.
Kaleido Diffusion: Improving Conditional Diffusion Models with Autoregressive Latent Modeling
·2911 words·14 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 Apple
Kaleido Diffusion boosts the diversity of images generated by diffusion models without sacrificing quality, using autoregressive latent modeling to add more control and interpretability to the image g…
Just Add $100 More: Augmenting Pseudo-LiDAR Point Cloud for Resolving Class-imbalance Problem
·4026 words·19 mins· loading · loading
AI Generated Computer Vision Object Detection 🏢 Korea University
Boost 3D object detection accuracy by augmenting pseudo-LiDAR point clouds!
Jointly Modeling Inter- & Intra-Modality Dependencies for Multi-modal Learning
·1862 words·9 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 Courant Institute of Mathematical Sciences
I2M2: A novel framework revolutionizes multi-modal learning by jointly modeling inter- and intra-modality dependencies, achieving superior performance across diverse real-world datasets.
John Ellipsoids via Lazy Updates
·311 words·2 mins· loading · loading
AI Theory Optimization 🏢 Carnegie Mellon University
Faster John ellipsoid computation achieved via lazy updates and fast matrix multiplication, improving efficiency and enabling low-space streaming algorithms.
JiuZhang3.0: Efficiently Improving Mathematical Reasoning by Training Small Data Synthesis Models
·1722 words·9 mins· loading · loading
Natural Language Processing Large Language Models 🏢 School of Information, Renmin University of China
JiuZhang3.0 efficiently enhances LLMs’ mathematical reasoning by training a small model to synthesize high-quality training data, drastically reducing costs.
Jailbreaking Large Language Models Against Moderation Guardrails via Cipher Characters
·2559 words·13 mins· loading · loading
Natural Language Processing Large Language Models 🏢 School of Information Sciences, University of Illinois at Urbana-Champaign
New benchmark and jailbreak method exposes vulnerabilities of LLM moderation, achieving significantly higher success rates than existing methods.
IWBVT: Instance Weighting-based Bias-Variance Trade-off for Crowdsourcing
·1247 words·6 mins· loading · loading
AI Applications Education 🏢 China University of Geosciences
IWBVT: A novel instance weighting approach significantly improves model quality in crowdsourcing by mitigating the impact of intractable instances and achieving a bias-variance trade-off.
iVideoGPT: Interactive VideoGPTs are Scalable World Models
·3466 words·17 mins· loading · loading
AI Applications Robotics 🏢 Tsinghua University
iVideoGPT: A scalable, interactive world model trained on millions of human & robot manipulation videos, enabling efficient video prediction and model-based reinforcement learning.
Iteratively Refined Early Interaction Alignment for Subgraph Matching based Graph Retrieval
·3157 words·15 mins· loading · loading
Machine Learning Deep Learning 🏢 UC San Diego
IsoNet++ iteratively refines subgraph matching via early interaction GNNs and node-pair partner interactions, significantly boosting graph retrieval accuracy.
Iteratively Refined Behavior Regularization for Offline Reinforcement Learning
·2346 words·12 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Shanxi University
Iteratively Refined Behavior Regularization boosts offline reinforcement learning by iteratively refining the reference policy, ensuring robust and effective control policy learning.
Iterative Reasoning Preference Optimization
·1561 words·8 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Meta FAIR
Iterative Reasoning Preference Optimization boosts large language model reasoning by iteratively refining preferences between generated reasoning steps, achieving significant accuracy gains on benchma…
Iterative Methods via Locally Evolving Set Process
·3065 words·15 mins· loading · loading
AI Theory Optimization 🏢 Fudan University
This paper proposes a novel framework, the locally evolving set process, to develop faster localized iterative methods for solving large-scale graph problems, achieving significant speedup over existi…
Iteration Head: A Mechanistic Study of Chain-of-Thought
·2483 words·12 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Meta AI
Researchers reveal how Chain-of-Thought reasoning emerges in transformers via specialized ‘iteration heads’, improving LLM performance and offering insights into mechanistic interpretability.
Is Value Learning Really the Main Bottleneck in Offline RL?
·2601 words·13 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 UC Berkeley
Offline RL’s performance often lags behind imitation learning, but this paper reveals that policy learning and generalization, not value function learning, are often the main bottlenecks.
Is the MMI Criterion Necessary for Interpretability? Degenerating Non-causal Features to Plain Noise for Self-Rationalization
·1904 words·9 mins· loading · loading
Natural Language Processing Text Classification 🏢 Huazhong University of Science and Technology
New criterion maximizes remaining discrepancy after rationale removal, treating spurious features as noise, improving rationale extraction.
Is Score Matching Suitable for Estimating Point Processes?
·1651 words·8 mins· loading · loading
AI Theory Optimization 🏢 Center for Applied Statistics and School of Statistics, Renmin University of China
Weighted score matching offers a consistent, efficient solution for estimating parameters in point processes, overcoming the limitations of previous methods.
Is Programming by Example solved by LLMs?
·2523 words·12 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Cornell University
Large Language Models (LLMs) surprisingly improve the challenging task of Programming by Example (PBE) when fine-tuned on problem-specific data, outperforming classic symbolic methods and even surpass…