Posters
2024
OccamLLM: Fast and Exact Language Model Arithmetic in a Single Step
·2170 words·11 mins·
loading
·
loading
Natural Language Processing
Large Language Models
π’ MIT
OccamLLM: LLMs now perform accurate arithmetic in a single step!
Object segmentation from common fate: Motion energy processing enables human-like zero-shot generalization to random dot stimuli
·2181 words·11 mins·
loading
·
loading
Computer Vision
Image Segmentation
π’ University of TΓΌbingen
Neuroscience-inspired motion energy processing enables human-like zero-shot generalization in figure-ground segmentation, outperforming deep learning models on random dot stimuli.
OASIS: Conditional Distribution Shaping for Offline Safe Reinforcement Learning
·2351 words·12 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
π’ Carnegie Mellon University
OASIS, a novel data-centric approach, shapes offline data distributions toward safer, higher-reward policies using a conditional diffusion model, outperforming existing offline safe RL methods.
NVRC: Neural Video Representation Compression
·1996 words·10 mins·
loading
·
loading
Computer Vision
Video Understanding
π’ Visual Information Lab, University of Bristol, UK
NVRC: A novel end-to-end neural video codec achieves 23% coding gain over VVC VTM by optimizing representation compression.
Nuclear Norm Regularization for Deep Learning
·1763 words·9 mins·
loading
·
loading
Machine Learning
Deep Learning
π’ MIT
This paper presents a novel, efficient method for Jacobian nuclear norm regularization in deep learning, replacing computationally expensive SVDs with equivalent Frobenius norm computations, thereby e…
Novel Object Synthesis via Adaptive Text-Image Harmony
·4696 words·23 mins·
loading
·
loading
AI Generated
Multimodal Learning
Vision-Language Models
π’ School of Computer Science and Engineering, Nanjing University of Science and Technology
Researchers created a novel object synthesis method, Adaptive Text-Image Harmony (ATIH), that harmoniously blends image and text inputs to generate creative, composite objects.
Not so griddy: Internal representations of RNNs path integrating more than one agent
·2491 words·12 mins·
loading
·
loading
AI Theory
Representation Learning
π’ Johns Hopkins Applied Physics Laboratory
RNNs trained on dual-agent path integration develop distinct internal representations compared to single-agent models, exhibiting weaker grid cell responses and enhanced border/band cell activity, wit…
Not Just Object, But State: Compositional Incremental Learning without Forgetting
·2423 words·12 mins·
loading
·
loading
Computer Vision
Image Classification
π’ Dalian University of Technology
CompILer: A novel prompt-based incremental learner mastering state-object compositions without forgetting, achieving state-of-the-art performance.
Normalization Layer Per-Example Gradients are Sufficient to Predict Gradient Noise Scale in Transformers
·2502 words·12 mins·
loading
·
loading
Natural Language Processing
Large Language Models
π’ Cerebras Systems
By cleverly integrating per-example gradient norm calculations during the backward pass of LayerNorm layers, this research enables efficient and accurate gradient noise scale estimation in Transformer…
Normalization and effective learning rates in reinforcement learning
·2714 words·13 mins·
loading
·
loading
Machine Learning
Reinforcement Learning
π’ Google DeepMind
Normalize-and-Project (NaP) boosts reinforcement learning by stabilizing layer normalization, preventing plasticity loss, and enabling effective learning rate control.
Normal-GS: 3D Gaussian Splatting with Normal-Involved Rendering
·2264 words·11 mins·
loading
·
loading
Computer Vision
3D Vision
π’ Monash University
Normal-GS improves 3D Gaussian Splatting by integrating normal vectors into the rendering pipeline, achieving near state-of-the-art visual quality with accurate surface normals in real-time.
Nonstationary Sparse Spectral Permanental Process
·2196 words·11 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
π’ Center for Applied Statistics and School of Statistics, Renmin University of China
Nonstationary Sparse Spectral Permanental Process (NSSPP) enhances point process modeling by using sparse spectral representations, enabling flexible, efficient, nonstationary kernel learning.
Nonparametric Instrumental Variable Regression through Stochastic Approximate Gradients
·1348 words·7 mins·
loading
·
loading
AI Theory
Causality
π’ Columbia University
SAGD-IV: a novel functional stochastic gradient descent algorithm for stable nonparametric instrumental variable regression, excelling in handling binary outcomes and various loss functions.
Nonparametric Evaluation of Noisy ICA Solutions
·3684 words·18 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
π’ Department of Computer Science, UT Austin
Adaptive algorithm selection for noisy ICA is achieved via a novel nonparametric independence score, improving accuracy and efficiency.
Nonparametric Classification on Low Dimensional Manifolds using Overparameterized Convolutional Residual Networks
·1457 words·7 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
π’ UC Santa Barbara
Overparameterized ConvResNets surprisingly excel at prediction; this study proves they efficiently learn smooth functions on low-dimensional manifolds, avoiding the curse of dimensionality.
Nonconvex Federated Learning on Compact Smooth Submanifolds With Heterogeneous Data
·1738 words·9 mins·
loading
·
loading
Machine Learning
Federated Learning
π’ KTH Royal Institute of Technology
This paper proposes a novel federated learning algorithm for nonconvex problems on compact smooth manifolds, achieving both computational and communication efficiency while mitigating client drift.
Non-Stationary Learning of Neural Networks with Automatic Soft Parameter Reset
·4994 words·24 mins·
loading
·
loading
AI Generated
Machine Learning
Reinforcement Learning
π’ Google DeepMind
AI models struggle with changing data; this paper introduces Soft Resets, a novel learning approach that uses an adaptive drift to gracefully guide parameters toward initialization, improving adaptabi…
Non-parametric classification via expand-and-sparsify representation
·1563 words·8 mins·
loading
·
loading
AI Generated
Machine Learning
Deep Learning
π’ Wichita State University
New non-parametric classifiers using expand-and-sparsify representation achieve minimax-optimal convergence, adapting to low-dimensional manifold structure.
Non-geodesically-convex optimization in the Wasserstein space
·332 words·2 mins·
loading
·
loading
AI Theory
Optimization
π’ Department of Computer Science, University of Helsinki
A novel semi Forward-Backward Euler scheme provides convergence guarantees for non-geodesically-convex optimization in Wasserstein space, advancing both sampling and optimization.
Non-Euclidean Mixture Model for Social Network Embedding
·2185 words·11 mins·
loading
·
loading
Machine Learning
Representation Learning
π’ UC Los Angeles
Non-Euclidean Mixture Model (NMM-GNN) outperforms existing methods by using spherical and hyperbolic spaces to model homophily and social influence in social network embedding, improving link predicti…