Skip to main content

Posters

2024

Learning and Transferring Sparse Contextual Bigrams with Linear Transformers
·1445 words·7 mins· loading · loading
Natural Language Processing Text Generation 🏢 Princeton University
Linear transformers efficiently learn sparse contextual bigrams by leveraging both in-context and global information, achieving polynomial sample complexity.
Learning an Actionable Discrete Diffusion Policy via Large-Scale Actionless Video Pre-Training
·2188 words·11 mins· loading · loading
AI Applications Robotics 🏢 Hong Kong University of Science and Technology
Actionable AI agents are trained efficiently via a novel framework, VPDD, which uses discrete diffusion to pre-train on massive human videos, and fine-tunes on limited robot data for superior multi-ta…
Learning a Single Neuron Robustly to Distributional Shifts and Adversarial Label Noise
·235 words·2 mins· loading · loading
AI Theory Robustness 🏢 University of Wisconsin-Madison
This work presents a computationally efficient algorithm that robustly learns a single neuron despite adversarial label noise and distributional shifts, providing provable approximation guarantees.
Learning 3D Garment Animation from Trajectories of A Piece of Cloth
·2097 words·10 mins· loading · loading
Computer Vision 3D Vision 🏢 Nanyang Technological University
Animates diverse garments realistically from a single cloth’s trajectory using a disentangled learning approach and Energy Unit Network (EUNet).
Learning 3D Equivariant Implicit Function with Patch-Level Pose-Invariant Representation
·2788 words·14 mins· loading · loading
Computer Vision 3D Vision 🏢 Xi'an Jiaotong University
3D surface reconstruction revolutionized: PEIF leverages patch-level pose-invariant representations and 3D patch-level equivariance for state-of-the-art accuracy, even with varied poses and datasets!
Learning 1D Causal Visual Representation with De-focus Attention Networks
·2168 words·11 mins· loading · loading
Multimodal Learning Vision-Language Models 🏢 Tsinghua University
De-focus Attention Networks achieve comparable performance to 2D non-causal models using 1D causal visual representation, solving the ‘over-focus’ issue in existing 1D causal vision models.
Learnability of high-dimensional targets by two-parameter models and gradient flow
·2386 words·12 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Skoltech
Two-parameter models can surprisingly learn high-dimensional targets with near-perfect accuracy using gradient flow, challenging the need for high-dimensional models.
Learnability Matters: Active Learning for Video Captioning
·2406 words·12 mins· loading · loading
Natural Language Processing Text Generation 🏢 Hangzhou Dianzi University
Active learning for video captioning is enhanced by a novel algorithm that prioritizes ’learnability’, diversity, and uncertainty to address annotation inconsistency.
Learn more, but bother less: parameter efficient continual learning
·2442 words·12 mins· loading · loading
Natural Language Processing Large Language Models 🏢 Pennsylvania State University
LB-CL: A novel parameter-efficient continual learning method for LLMs that boosts performance and reduces forgetting by leveraging parametric knowledge transfer and maintaining orthogonal low-rank sub…
LCM: Locally Constrained Compact Point Cloud Model for Masked Point Modeling
·2913 words·14 mins· loading · loading
AI Generated Computer Vision 3D Vision 🏢 Tsinghua University
LCM: a novel, locally constrained, compact point cloud model surpasses Transformer-based methods by significantly improving performance and efficiency in various downstream tasks.
LCGen: Mining in Low-Certainty Generation for View-consistent Text-to-3D
·2307 words·11 mins· loading · loading
Natural Language Processing Text Generation 🏢 Shanghai Engineering Research Center of AI & Robotics, Academy for Engineering & Technology, Fudan University
LCGen: A novel method for view-consistent text-to-3D generation, resolving the ‘Janus Problem’ by strategically using low-certainty priors to align viewpoints and optimize the generation process.
Layer-Adaptive State Pruning for Deep State Space Models
·2474 words·12 mins· loading · loading
Machine Learning Deep Learning 🏢 Department of Electrical Engineering, POSTECH
Layer-Adaptive STate pruning (LAST) optimizes deep state space models by efficiently reducing state dimensions, improving performance and scalability without retraining.
Latent Representation Matters: Human-like Sketches in One-shot Drawing Tasks
·3093 words·15 mins· loading · loading
Computer Vision Image Generation 🏢 Artificial and Natural Intelligence Toulouse Institute
AI now draws almost as well as humans, thanks to novel latent diffusion model regularizations that mimic human cognitive biases.
Latent Plan Transformer for Trajectory Abstraction: Planning as Latent Space Inference
·2609 words·13 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏢 UC Los Angeles
Latent Plan Transformer (LPT) solves long-term planning challenges in reinforcement learning by using latent variables to connect trajectory generation with final returns, achieving competitive result…
Latent Paraphrasing: Perturbation on Layers Improves Knowledge Injection in Language Models
·2300 words·11 mins· loading · loading
Natural Language Processing Large Language Models 🏢 KRAFTON
LaPael improves LLM knowledge injection by applying learned noise to early layers, enabling diverse and efficient knowledge updates without repeated external model usage.
Latent Neural Operator for Solving Forward and Inverse PDE Problems
·2797 words·14 mins· loading · loading
AI Theory Optimization 🏢 Institute of Automation, Chinese Academy of Sciences
Latent Neural Operator (LNO) dramatically improves solving PDEs by using a latent space, boosting accuracy and reducing computation costs.
Latent Learning Progress Drives Autonomous Goal Selection in Human Reinforcement Learning
·2787 words·14 mins· loading · loading
AI Generated Machine Learning Reinforcement Learning 🏢 UC Berkeley
Humans autonomously select goals based on both observed and latent learning progress, impacting goal-conditioned policy learning.
Latent Functional Maps: a spectral framework for representation alignment
·2758 words·13 mins· loading · loading
AI Generated Machine Learning Representation Learning 🏢 IST Austria
Latent Functional Maps (LFM) offers a novel spectral framework for comparing, aligning, and transferring neural network representations, boosting downstream task performance and interpretability.
Last-Iterate Global Convergence of Policy Gradients for Constrained Reinforcement Learning
·2971 words·14 mins· loading · loading
Machine Learning Reinforcement Learning 🏢 Politecnico Di Milano
New CRL algorithms guarantee global convergence, handle multiple constraints and various risk measures, improving safety and robustness in AI.
Last-Iterate Convergence for Generalized Frank-Wolfe in Monotone Variational Inequalities
·1879 words·9 mins· loading · loading
AI Generated AI Theory Optimization 🏢 Purdue IE
Generalized Frank-Wolfe algorithm achieves fast last-iterate convergence for constrained monotone variational inequalities, even with noisy data.